We’ve talked here in several issues about the “great resignation” that many companies are seeing as the pandemic starts to wane a bit. I think this is playing out differently in our world; there are definitely people leaving, or making new work arrangements, but it’s less en masse. Rather than having trouble with people leaving, instead, it seems to be more of an (even bigger) problem bringing people on.
On the jobs board there have been academic positions open since late winter, some that have been re-posted an uncomfortable number of times. That’s for manager positions; I’m also hearing from a lot of people that hiring ICs is getting almost impossibly difficult.
And it’s not hard to imagine why. In the last couple weeks I’ve spoken with managers at tech companies (mainly through online communities) who were actively hiring ICs and the total compensation they were describing was routinely a comfortable 2.5-3x higher than I can pay. And my organization pays quite well for academia. These weren’t some horrible soul-rending ad-tech jobs aimed at increasing click-through-rates, either, but interesting, challenging, and high-impact infrastructure or software or operations jobs relevant to research computing.
I mentioned in #90 that teams needed to be ready to compete, and they do; not just for work but for workers. Having a well-defined specialty and being known for something helps. And in #81 I mentioned that too often we underplay our strengths as employers - stability of employment with variety of work; campus or remote, either of which is attractive to a population; work that we can be proud of.
Collectively, we’re going to need to up our game when it comes to hiring. Some things change slowly with time and a lot of collective effort, such as developing career paths, improving compensation, etc. But other things are entirely within our control. Writing job ads that raise the frankly abysmal bar set in our industry; spending time actively recruiting; being clear on what the benefits of our team is and what our team’s goals are; having a well-thought through, well-communicated, non- made-up-on-the fly interview process; making sure team members get the support they need, and opportunities to shine in venues they care about. These are things we don’t need anyone’s permission to do, don’t require HR sign-off, and by and large don’t require funding approval.
Hiring’s hard and isn’t going to get easier any time soon. There are a lot of exciting and well-paid challenges outside of our industry. We can compete! But we need to try.
On to the roundup:
The effects of remote work on collaboration among information workers - Longqi Yang et al., Nature Human Behaviour
This is a peer-reviewed paper covering some data reported earlier in blog posts; Microsoft looking at internal messaging and calendaring traffic of its US employees over the first six months of 2020. They found that, internally, the size of communications networks fell; groups became more siloed, and as a result presumably it is harder to communicate across different divisions of an organization.
This is consistent with what we found in our team. We were in a slightly unusual situation - we had a multi-site institution. At least at our site, I find that the collaboration’s communications is actually much better than before the pandemic - we put a lot of work into that - but there’s no free lunch, and my communication with other teams at my home institution is relatively poor.
By and large, I think this works out ok for many research computing teams; we’re pretty self-contained, and consult with researchers on a project-by-project basis for focussed periods of time. But it does point to areas where managers need to work - making sure there are bridges built to other internal teams we need to interact with (IT? VPR?) and putting extra work into communicating messages that we want heard widely like successes or changes.
How to coach an employee who is performing well - Claire Lew, Know Your Team
Researchers not taking enough time for career development - Sophie Inge, Research Professional News (paywalled)
As managers, the tendency is to focus on problems, and on team members who aren’t performing well in their current tasks. As a result, team members who are capable and already growing tend to suffer from benign neglect; but this isn’t fair to them, and it’s a missed opportunity for the manager. Someone who’s already shown they can grow under conditions of benign neglect can grow faster with some help. What’s more, if you don’t offer that help, they may reasonably start looking for jobs where their growth and career development is actively encouraged - wouldn’t you?
Lew walks us through the steps of making sure someone who’s doing well and is already growing gets the coaching they deserve. As always, coaching isn’t about you telling someone how to do something better, but about giving them the opportunities and resources - and accountability - to grow their skills:
As with so much of management, this isn’t hard or complex! The difficulty is in the discipline to keep performing the practice. In #68 I shared the quarterly goal setting & review template I use; by checking in every three months, and understanding their goals in their larger career, as well as at this job, that makes it much easier to ensure that new opportunities and resources support them in their growth as well as in the work the team needs.
All of this is incredibly important; as Inge points out in a paywalled article, while researchers know that career development is vitally important, they find themselves “too busy” to improve their own skills and thus productivity. That attitude spreads outwards to researcher-adjacent staff. We’d all agree that putting time and effort into growth and development is vital, and yet too often it doesn’t get done. One of our biggest roles as a manager is to make sure our team members (and ourselves) do take the time to grow our skills, so we can take on more ambitious efforts and be more satisfied in our work.
Why the best way to hire is incredibly boring - Laszlo Bock, humu
Again - management isn’t complicated, it doesn’t require profound insights into the human condition or the right personality type - it consists of nothing more than the discipline of attending to the details.
Bock despairs of the “one weird interview question” genre of articles that promises to give you that profound insight into a candidate without having to do any real work.
But hiring doesn’t work like that. When hiring a human being to do a job for the team and grow the team, you need to find out if they can do the job, and if they’ll grow the team in a way that the team would benefit from. As Bock points out, the way to do that is exactly what you’d expect:
It’s boring, it’s mundane, it’s a lot of work, and there isn’t another good general option.
We recently went through and have nearly completed hiring a technical project manager (I can’t even tell you how excited I am by that prospect), so this is all fresh in my mind. This is our first TPjM hire, so we spent a lot of time putting together interview materials to determine what we needed, what a new team member would need to do to demonstrate success in the role, what the criteria are, and then how to get that “work sample” - in this case, a “project planning meeting” interview. in addition to the behavioural questions interview (where we provided the questions ahead of time, along with the scenario). It was a lot of work, it greatly clarified within the team what we needed, it greatly clarified for the candidate what they were proposing getting into, and it basically sets up the goals for the first six months of the candidate. It improves our, and their, chances of success.
The one place I’d disagree with Bock is about averaging scores. You don’t want to hire someone that, on average, the team wants to work with.
You absolutely have to weigh all the input, and use that to inform your decision. But I’d suggest two things.
First, any well-justified objection to a candidate based on the agreed-upon criteria should be pretty darn close to disqualifying. Second, the “decision” can’t be made by an artificially dispassionate scoring system, as comfortable as that would be. The decision has to ultimately be your responsibility, the hiring manger - for the very simple reason that if it goes horribly wrong and they have to be fired, that responsibility will fall on you and no one else.
Toward Modern Fortran Tooling and a Thriving Developer Community - Milan Curcica et al., arXiv:2109.07382
Back in #56 we pointed out a blog post celebrating one year of the new Fortran-lang community; in this preprint released this week, Curcica et al go into more details about the creation of a community organized around GitHub, the development of a standard library (a short 64 years after the introduction of the language), a package manager, and a community-curated website where there had previously only been quarterly standards group meetings.
It’s an encouraging reminder that communities want to form, and that a small number of people willing to put the time in can catalyze a surprising amount of effort (here, over 60 people) focused on improving the state of affairs within their niche.
Loading NumPy arrays from disk: mmap() vs. Zarr/HDF5 - Itamar Turner-Trauring
This article is written with Python examples (and Zarr in particular is a python package), but Turner-Trauring’s basic discussion of mmap() as a tool for accessing large files vs specialized libraries is quite general.
mmap() isn’t magic - and it still involves a lot of the mechanisms implemented for file systems. It can be a very quick way to get started, and if you have genuinely random access it may be pretty close to optimal (though there’s some tuning you can do still) but if you know something about the structure with which you’ll be accessing your data, you can often do better - particularly if you can take advantage of compression.
Here are two very pragmatic articles about documentation for developers. The common theme here is that the most important thing about changelings and class documentation is that they exist. And the easiest way to ensure that’s done is to have simple, clear, and non-onerous expectations of what they should look like.
Rocha and Lui, covering change logs, specify:
Gulbrandsen, on writing class documentation, spells out the simple formula he used for Qt:
Good writers will do it differently, but the goal here is to write something good, not to be a good writer. The best is the enemy of the good.
The Data Product Manager - Rifat Majumder, Locally Optimistic
I’ve been beating the drum for some time that research computing teams need to start thinking less in terms of projects and more in terms of products. So far, the only place I’ve seen that happen are in areas very close to data, with data products themselves, or data processing pipelines as products - and even so I’ve only seen 4/224 academic jobs on the job board with that title.
In this article, Majumder describes the benefits of treating curated data collections as a product, including:
The article is written in the context of internal data products for use within a large company, but it works just as well with management of large data resources for research communities with some of the words changed.
Beard provides a LaTeX template for Gebru et al’s suggested “Datasheets for Data sets”, a human readable high level description of a dataset - not a data dictionary, but describing the reason the data set exists, how data was collected, what preprocessing/cleaning/labeling was done if any, how or if maintenance will be done, what uses the dataset has been put to, and more.
Napkin Problem 16: When To Write a Simulator - Simon Eskildsen
Eskildsen suggests, rather than rely on intuition (or back-of-the-napkin envelopes), we should
Simulate anything that involves more than one probability, probabilities over time, or queues.
When we’re thinking about batch queueing system parameters, or sharding, or load balancing, or anything else, our systems usefulness depends on what we choose, but we too often choose it based on trial-and-error or rules of thumb. We work in the sciences, and writing simple simulations of our parameters to guide our decisions isn’t normally very difficult; it can expose corner cases that you mightn’t otherwise consider; and it lets you run simple what-if calculations.
Assessing the Use Cases of Persistent Memory in High-Performance Scientific Computing - Yehonatan Fridman et al., arXiv:2109.02166
In this paper, Fridman et al use Optane persistent memory devices as 1) memory, 2) swap, or 3) as disks for checkpoint restart. They take a look at microbenchmarks, several of the NAS Parallel Benchmarks and Polybench benchmarks. There are definitely benchmarks for which using the Optane in memory mode offers only 50% performance drop but for much larger problem sizes, for more or less free, as in the plots shown below:
But there’s benchmarks for which the drop is much worse. Using the optane as swap gave pretty uniformly poor results. It worked quite well for crash/restart. The code the team used is available on github.
In general, I think we’re going to need to try harder than just using these nonvolatile devices as memory or as a drive to get sort of performance we’d like from codes that use them. There was a brief window in the 70s-90s where there was active development of external-memory/out-of-core algorithms (and then in the 2000s, cache oblivious algorithms) to make use of multi-stage memory hierarchies. As we move to NVMe, and to RDMA over PCIe busses, I suspect we’re going to have to dust those off. One interesting very long review is here (PDF).
If you’re thinking of starting up programs around reliability for your systems, nginx has a free (in exchange for your email address) O’Reilly ebook, 97 Things Every SRE Should Know. It covers the basic mental models of tech-industry SRE basics, how to get started, then increasingly advanced processes (ranging from having a single person being the system reliability engineering “team” to 10-100).
SRE sounds kind of exotic and mysterious and advanced to those of us in the research computing sector, but it’s not magic, nor is it arcane - it’s a bunch of practices centred around being extremely clear about what “reliability” means in your context, defining metrics so you know if you’re being reliable enough (or even more reliable than you need to be!), and then being deliberate about maintaining reliability while making changes.
Many of the 97 (98, really) 1-2 page essays likely won’t be applicable to your particular situation, but it’s a good crash course in the language and models of SRE and will help you get a foothold in the huge SRE literature out there.
Deploying Kubernetes Based HPC Clusters in a Multi-Cloud Environment - Daniel Gruber，Burak Yenier, Wolfgang Gentzsch, HPCWire
Owl Job Scheduler - submit user jobs to a Kubernetes cluster from anywhere - Eduardo Gonzalez
“Let’s use Kubernetes!” Now you have 8 problems - Itamar Turner-Trauring
I’m pretty skeptical of using Kubernetes for running traditional HPC jobs - Kubernetes on the one hand and (say) Slurm or SGE something on the other serve two completely different use cases, and something that’s well served by category A is not likely to be a good match for category B.
However. HPC schedulers are so well optimized because they choose a simple problem - single “rectangular” jobs, N nodes x M hours - which aren’t a good match for all job types. Across jobs they have widely varying capabilities (mostly ranging from “meh” to “ugh”) for handling complex DAGs of tasks needed by modern workflows. Further, despite noble efforts with DRMAA v1 and v2, API-driven programatic control of schedulers, particularly across systems, is still really poor. Finally, if you are running in an environment with elastic compute resources, the extremely sophisticated HPC schedulers that are used to get the highest utilization to out of fixed static compute resources aren’t nearly as valuable.
So there’s a lot of interest in targeting Kubernetes as a single uniform interface to resources wherever they may be. Again, I think it’s a weird match for HPC - simultaneously overkill and lacking some key functionality - but it’s about as close as we have to a standard API for compute resources right now, so here we are.
In the first set of articles, Gruber, Yenier, and Gentzsch talk about their significant experience using Kubernetes as part of the Ubercloud effort to provide cloud-based HPC for engineering simulations for a variety of customers, culminating in a study of 3000 simulations on spot instances, each within their own ephemeral kubernetes cluster (!) on GKE, calling services from AWS and Azure. In their telling, the extra overhead cost involving GKE was significantly overcome by the savings from nicely handling spot instance preemption.
On the second page, Gonzalez describes Owl job scheduler for jobs on Kubernetes - testing locally on Docker Desktop, or running on production on GKE or other Kubernes clusters. Demonstrated are data science jobs using Python + Dask, pipelines on single-cell data, and Jupyter notebooks.
On the “con” side, Turner-Trauring gives the conventional take on Kubernetes, and one I share, which is that it’s a huge thing to take on, it’s enormously complex, and you’d do well to make absolutely sure it’s right for your team’s needs before starting with it. A recent article talked about one team moving off managed kubernetes (GKE) to nomad for a single site specifically because of ease of handling batch jobs.
Software Sustainability Fellowships - £3,000/15 month, primarily for UK-affiliated people, due date 31 Oct
From the site:
This funding can be used for any activities that meet both the Fellow’s and the Institute’s goals, such as travel to workshops, running training events such as Software Carpentry, Data Carpentry or Library Carpentry, nurturing or contributing to communities of practice, collaborating with other Fellows, or for any other activities that relate to improving computational practice or policy.
SeptembRSE Open Data Workshop - 28 Sept
This AWS/SeptembRSE event focused on hackathon for teams of up to five for running reproducible analyses on one of the AWS-hosted open datasets:
Join us for a distributed, hybrid workshop on using Amazon Web Services (AWS) to run reproducible analyses from the registry of open data. Inspired by the reprohack model, participating teams will recreate example workflows, deploy infrastructure, and run analysis.
NVIDIA GTC - 8-11 Nov, Virtual - Free registration for talks, full-day workshops $99
NVIDIA’s huge GTC is open for registration, with talks on methods and applications for genomics, image processing for astronomy, medicine, and other areas, natural language processing, molecular dynamics simulations, DPUs for supercomputing, infiniband, the current NVIDIA HPC stack, GPUs fluid simulations, and more.
PRACE Winter School 2021-Converging HPC Infrastructure & Methodologies - 7-9 December, Free, Registration limited (due by 15 Oct) and priority given to regional attendees
The Inter-University Computation Center (IUCC) in Israel is proud to host the PRACE 2021 Winter School which will feature a rich introductory workshop on how to use PRACE resources as well as trends in future software and data ecosystems for scientific discovery.
Computer games for education have been around since the beginning - here’s the story of 1964’s The Sumerian Game, complete with teletype, slide projector and narration on cassette. I played the 1978 version of this, typed in from Basic Microcomputer Games.
It turns out a couple approximate math instructions (approximate single precision reciprocal and square root) vary slightly from AMD to Intel, which is a problem if you’re emulating. Here’s how the team working on the reverse debugging package rr handled it.
Not to be outdone, Racket (a variant of scheme) has a GUI framework too, they want you to know.
A couple make-golang-go-faster tools - a tutorial on go profiling, tracing, and observability, and go assembly language libraries for a variety of platforms for sortedsets, byteswapping, qsort, etc with native go fallbacks.
In praise of awk, a too-often unsung hero.
Using AWS ECS and Fargate to run containers on a raspberry pi at home. Here the author uses a tool called inlets to handle traffic back through a home NAT with dynamic IP, but you could use something like tailscale or other fireguard VPN as well I think.
Functionality of Alpline Linux I didn’t know about - individual service-level network isolation with virtual routing and forwarding (VRFs).
Case studies (from an Oracle employee) of using Oracle’s always-free tiers of services for real services, including a simple data-driven website or development/test instances.
Newish programming language Zig is taking a very different approach to Rust when it comes to working with C/C++ code; as opposed to the “Rewrite It In Rust” mantra, Zig is making interoperability with followed by incremental additions/replacements to C/C++ code with Zig code a priority.
The newish buildkit tool, most often used with Docker, also works with podman.
Distinguishing between OS package managers and e.g. programming language package managers - closed universe vs open universe.
And that’s it for another week. Let me know what you thought, or if you have anything you’d like to share about the newsletter or management. Just email me or reply to this newsletter if you get it in your inbox.
Have a great weekend, and good luck in the coming week with your research computing team,
About This Newsletter
Research computing - the intertwined streams of software development, systems, data management and analysis - is much more than technology. It’s teams, it’s communities, it’s product management - it’s people. It’s also one of the most important ways we can be supporting science, scholarship, and R&D today.
So research computing teams are too important to research to be managed poorly. But no one teaches us how to be effective managers and leaders in academia. We have an advantage, though - working in research collaborations have taught us the advanced management skills, but not the basics.
This newsletter focusses on providing new and experienced research computing and data managers the tools they need to be good managers without the stress, and to help their teams achieve great results and grow their careers.
This week’s new-listing highlights are below; the full listing of 171 jobs is, as ever, available on the job board.
Senior Systems Manager - British Trust for Ornithology, Thetford UK
For almost 90 years, BTO has been, and remains, an organisation built around data. The development of computer technology and the advent of the Internet were embraced by the organisation. However, the pace of change continues to be rapid and BTO is dedicated to keeping at the forefront of information systems. We are now recruiting a Senior Systems Manager, who would have overall responsibility for providing computer infrastructure, software packages and support to about 130 staff, based variously at our HQ (The Nunnery, in Thetford), our satellite offices, and working from home; and to support online access to BTO web applications by our supporters and the wider public. Line management of the Computing Support Officer. Helping develop a DevOps culture within the team by provision of a high quality computing service for BTO volunteers and the public by maintaining and developing the infrastructure needed to support BTO’s on-line applications.
Director of Bioinformatics Engineering - Genomics England, London UK
This is an exciting opportunity for an expert in bioinformatics and software engineering to join an amazing team. You’ll be collaborating with many parts of the business, performing a range of senior functions including team leadership, external relationship management (e.g. with GA4GH), innovation, quality and strategy. In this leading role, you’ll drive the adoption of best-in-class bioinformatics solutions by working with architects and engineers across the organisation and ensuring that what we build is grounded in solid bioinformatics and genomics. Importantly, You will be our head coach for bioinformaticians and bioinformatics engineers, providing them with mentoring, career development, and training opportunities.
Principal Biostatistician - IQVIA, Kirkland QC CA
Develop statistical methods sections of protocols and review case report forms (CRFs). Prepare analysis plans and write specifications for analysis files, tables, and figures. Communicate with clients regarding study protocol or statistical analysis issues as they arise. Communicate with study team members regarding study execution as it relates to timelines, data quality, and interpretation of results. Interpret analyses and write statistical sections of study reports. Accountable for controlling costs and maximizing revenue recognition. Provide training, guidance and mentorship to lower level and new staff.
Senior Research Associate (Research Software Engineer) - University of Bristol, Bristol UK
he OpenGWAS project (https://gwas.mrcieu.ac.uk/) aims to i) collect and harmonise results from genome-wide association studies; ii) make these publicly accessible via web services and downloads; iii) connect these data sources to a range of statistical tools in R and python; and iv) automate systematic analysis across the surface of data and tools as the resource grows. The database currently comprises over 180 billion genetic associations, is growing rapidly and serves thousands of users. You will lead the development and maintenance of R packages that connect a range of genetic summary data sources with analytical tools, and which are widely used for enabling rapid and reproducible analyses at large scales. This work will occur alongside the development of tools for Python users, and require collaboration in building common data models.
Manager, Data Science and Analytics - CircleCI, Toronto ON CA
CircleCI is seeking a Manager, Data Science and Analytics to lead our data team. In this role, you will help our product organization and teams understand their customer better through insights and analytics. You will build a team that delivers impactful insights through storytelling in a hyperscale environment.
Manager, Data Science - AltaML, Edmonton AB CA
As a technical leader in projects, you will lead teams to develop AI/ML solutions utilizing Data Mining and Machine Learning techniques. You will work with multiple tools and gain experience in many AI/ML and data processing fields. You will ensure that ML developers follow our rigorous ML development processes and will work on improving them. You will also be accountable for the quality of assurance processes in the projects you lead. You will be given resources and mentorship to help with your growth and career advancement.
Senior Manager, Data Science - GitHub, Remote CA or USA
Your job will be to lead both the Developer Cloud and Security pillars within Data Science at GitHub and partner with Data Engineering teams in these respective pillars. The Data Science team is highly distributed and the right candidate will thrive in an environment of asynchronous communication. We expect you to have excellent written communication skills and be able to create working relationships with coworkers in locations around the globe. In this role also you will lead, provide guidance, support and advice to your team.
Senior Exascale Software Engineer (Fixed Term) - University of Cambridge, Cambridge UK
The Senior Exascale Software Engineers play a key role in the COEL, developing and optimising new and existing complex scientific applications to exploit state of the art supercomputing technologies which enable the research agendas of the University’s and the wider international research community. Significant experience will be required for the senior roles in addition to supervisory and mentoring skills.
Principal Exascale Software Engineer (Fixed Term) - University of Cambridge, Cambridge UK
The Principal Exascale Software Engineer will manage a small team of highly technical research software engineers on multiple projects to develop and optimise advanced scientific applications for the next generation of supercomputing technologies. The successful candidate will work with closely with the technical programme manager, key academic collaborators and external stakeholders to define and assure successful delivery of the projects.
Quantum Computing Research Lead - 1QBit, Vancouver BC CA
As a research lead, you will have the opportunity to work with the executive team to define exciting long term research roadmaps and direction, help grow the research team and collaborate with world-leading scientists at our partner research organizations, as well as a pioneering team of peers. You will be able to get your hands on cutting-edge quantum computing platforms, have a front-row seat in ground-breaking developments in the field, and drive your team to contribute to it. You are encouraged to think creatively and help define our research path toward the application of NISQ to practical chemistry problems.
EMDB Resource Manager - EMBL, Hinxton UK
This is an exciting opportunity for an expert in the field of electron microscopy to lead the Electron Microscopy Data Bank (EMDB). the EMDB Resource Manager will lead a small team (presently 7 staff members, a mixture of structural biologists and computational scientists) and be responsible for leading EMDB’s archival, curation, integration, visualisation and dissemination activities, its production processes (weekly release preparation, updates of databases, websites, ftp areas, etc.), and the active involvement of the team in the wwPDB-related work (annotation of EMDB data, planning and implementing updates of the unified wwPDB deposition system OneDep, validation software, data model, etc.). Currently, the EMDB activities are funded by EMBL-EBI and two grants from the Wellcome Trust. We expect the EMDB Coordinator to play an important role in attracting future external funding.
Software Architect - University of Manchester, Manchester UK
The Digital Health Software team within the Centre for Health Informatics at the University of Manchester is a team of software engineers, designers, project managers, digital health researchers. We have successfully developed a number of innovative digital health technologies that aim to improve patient care across multiple health conditions. ClinTouch (https://www.herc.ac.uk/research_project/clintouch/) is one of our success stories and is a mobile app that enables personalised clinical monitoring and supports self-management of symptoms for people with serious mental illness. CFHealthHub (https://www.herc.ac.uk/research_project/cfhealthhub/) is another successful project and provides a digital platform (web and mobile) to support adherence to treatment in people with Cystic Fibrosis. Details of the projects that we have and are currently working on can be found here: http://www.herc.ac.uk/our-research/research-projects/
Faculty and Research Support Manager - University of Texas Rio Grande Valley, Edinburg TX USA
Manages the day-to-day operation, security, system administration and maintenance of high-performance computing (HPC) systems to accomplish the research objectives of faculty and students at the University. Provides user support, documentation, and consulting services to ensure the effective use of research computation resources for faculty members and students.
Scientific Computing Sr Manager - Fred Hutchinson Cancer Research Center, Seattle WA USA
The Scientific Computing Senior Manager will direct and oversee all facets of scientific computing including managing a highly capable engineer team, engineering functions such as design, development, installation, and maintenance of hardware and software, and customer service and support, for the organization. The senior manager will create and oversee partnerships with faculty and stakeholders in the scientific community to test and integrate new analysis pipelines, determine innovative technologies and software to support their research, evangelize new and current solutions, and foster technology and data storage best practices. The senior manager will participate in strategic planning and align projects and resources for implementation in partnership with the Project Management Office and Business Operations, and other departments.
Clinical Science Team Lead - Bristol-Myers Squibb, San Diego CA USA
Leads the Clinical Science team in the planning and execution of clinical trial activities to support the Clinical Development Plan for the disease/indication team. Manages and supervises a team of senior and junior Clinical Scientists to execute trial level activities for multiple trials. Ensures consistency across studies and provide oversight of all clinical studies within assigned area. May serve as a Clinical Trial Lead or Co-Lead as needed. May represent the Clinical Scientist function in various process improvement initiatives and/or cross functional activities
Senior Data Team Lead, Real World Evidence - IQVIA, Cambridge MA USA
• Serve as primary point of contact for customer on data management deliverables • manage multiple large studies or program of studies (possibly global) Provide leadership and senior support to DTLs on multiple large global studies/programme of global studies Provide leadership and senior support to DTLs on multiple large global studies/programme of global studies. • Provide project management expertise working with customer data managers, key decision makers, and internal team members to manage continuous process improvements, issue escalation, workload projections, and provide technical expertise • Lead direct negotiations with customer, e.g., timelines, financial, process, resources. • Maintain strong customer relationships
Head of Research Computing - Norwich Bioscience Institute Partnership, Norwich UK
We have an exciting opportunity for a strategic High-Performance Computing (HPC) leader to take accountability for the future development of research computing in one of the UK’s foremost research settings. The Head of Research Computing will set the technology strategy and service model for mission-critical research IT services at NBI. The Norwich Bioscience Institutes (NBI) are a cluster of internationally renowned research organisations. The institutes work together to tackle the major challenges facing us all in the 21st Century – the sustainability of our environment; our food supplies and healthy ageing. Across NBI, there are over 1,000 scientists working to find realistic and practical solutions to these challenges. We then provide the infrastructure and support to translate their discoveries into commercially-successful businesses.
Associate Director, Research Computing - Simon Fraser University, Burnaby BC CA
The Associate Director, Research Computing leads a team dedicated to gathering requirements, analyzing, evaluating, implementing, validating, and supporting solutions for advanced research computing to support SFU’s strategic goals. Working in partnership with members of the university community, the Associate Director plans, coordinates, and manages research computing solutions; the deployment of national and university specific services; and the allocation of team resources to successfully complete projects that align with the university’s priorities.
Research Support Senior Manager - Jisc, various UK
The e-Infrastructure Strategy Group implements strategies for supporting the UK’s national and global research, overseeing work on how digital infrastructures can support and enhance aspects of the research lifecycle - from discovery of information and data, to data analysis and manipulation, and collaboration and research impact and dissemination. We’re ideally looking for someone with Experience of identifying issues impacting the research community, specifically within the use of digital tools and services, and how to address them strategically; Experience of working in a technical research support role, or within a research or University environment where you have managed an IT infrastructure team; Significant experience in analysing complex requirements and resolving multifaceted problems; Excellent skills in network development and emerging technologies
Principal Technical Program Manager, Quantum Computing - AWS, Pasadena CA USA
The Amazon Web Services (AWS) Center for Quantum Computing in Pasadena, CA, is looking to hire a Principal Technical Program Manager to coordinate and manage research and development activities across a diverse team consisting of scientists, hardware engineers/technicians, and software developers. We are committed to the growth and development of every member of the Center for Quantum Computing. You will receive career-growth-minded management and mentorship from a software and science team and also have the opportunity to participate in Amazon’s mentorship programs. You will work closely with quantum research scientists and have opportunities to learn about quantum computing technology and contribute to the development of scientific software for quantum computing at AWS.