Expression of Interest
We have an interdisciplinary team, and we are constantly in search of talent that can add to our team's skill set. Please submit this expression of interest if you are interested in working with us. We won’t scan applications on a regular basis, but we will reach out if there’s an open position we think you might be a good fit for.
Expression of Interest
You will join OpenMined, an open source community led by Andrew Trask (author of Grokking Deep Learning), as an expression of your interest in joining the OpenMined community, this is an opportunity to tell us about your background and skills. You can contribute in a way that leverages your industry experience, offers you an opportunity to grow new skills, while helping us solve the remote data access problem for the benefit of all.
Pro-Social Agent Engineer
Create a forkable virtual world filled with pro-social AI agents that help humans reach their goals while respecting the basic rights of meat.
Senior Software Engineer
If you want to democratize access to cutting edge AI, while averting potential harmful impacts on society, OpenMined is the place for you. We’re building PySyft, an open source software platform for doing just that, and we need you!
We are looking for a highly talented individual, with hard won industry experience, a track record of exceeding expectations, and who is ready to level up and do the most challenging and rewarding thing in their career so far.
Machine Learning Engineer
As an ML research engineer at Elicit, you will:
Compose together tens to thousands of calls to language models to accomplish tasks that we can't accomplish with a single call.
Curate datasets for finetuning models, e.g. for training models to extract policy conclusions from papers
Set up evaluation metrics that tell us what changes to our models or training setup are improvements
Scale up semantic search from a few thousand documents to 100k+ documents
Interpretability Researcher
EleutherAI is seeking talented and motivated individuals to join our Interpretability team to perform cutting edge research with large language and vision models. We aim to better understand the features learned by today’s deep neural networks, so we can better steer their behavior and inform the public and policy makers about their risks and benefits.
AI Safety and Security Research Engineer
As an AI Safety & Security Research Engineer, you'll advance the state of the art in AI safety and security while developing practical customer-facing tools and products. Your role involves developing novel methods for controlling, monitoring, testing, and analyzing foundation models, as well as building new models with a focus on scalable, real-world deployment. Staying abreast of the latest machine learning advancements is crucial, as you'll contribute to open model innovations and ensure our products remain at the forefront of AI technology. This position blends research with hands-on implementation, requiring both theoretical expertise and practical problem-solving skills to address complex challenges in AI safety and security.