AutoML and AI coding assistants are lowering the barrier for building models, but defining the right problem, understanding the business context, and translating statistical findings into decisions that organisations actually act on is still distinctly human work. Here is what the research says about the data scientist profession in 2026, and what you can do about it.
Get My Personalised Fossil ScoreFossil Score
62
Species
Velociraptor
AutoML and AI coding assistants are lowering the barrier for building models, but defining the right problem, understanding the business context, and translating statistical findings into decisions that organisations actually act on is still distinctly human work.
Task Automation Risk
42%
of current data scientist tasks are automatable with existing AI tools
Data science is experiencing a version of the same pressure affecting software engineering: AI tools can now write code, suggest model architectures, perform feature engineering, and explain statistical output in plain language. AutoML platforms (Google AutoML, H2O.ai, DataRobot) can build and evaluate models from labelled data without manual model selection. GitHub Copilot handles much of the Python boilerplate. This reduces the execution cost of standard modelling work and means that a smaller team or a business analyst with AI assistance can do what previously required a dedicated data scientist for some use cases. The 42% risk reflects this automation pressure on the routine modelling and code-writing work. What remains distinctly valuable: understanding which questions are worth asking of data in a specific business context; designing experiments to answer causal questions rather than correlational ones; explaining model outputs in ways that drive decisions rather than just reporting metrics; and handling the data quality, ethics, and governance questions that sit around every model deployment. Data scientists who develop business fluency — speaking in terms of revenue impact, operational trade-offs, and decision framing — alongside their technical skills are building the version of this role that AI doesn't replace.
Task Autopsy
🦕 Class A — At Risk Now
🦅 Class C — Protected
Your AI Toolkit
You don't need to learn all of these. Pick one, use it for a week, and see how it fits into your work. Most have free options so you can try before you commit.
Unified data and AI platform widely used at data-mature organisations — Delta Lake, MLflow experiment tracking, and SQL analytics in one platform; Databricks certification is increasingly recognised as a credential for data science roles
Try it ↗Experiment tracking and model management platform — logs training runs, hyperparameters, and metrics for comparison; standard tooling in organisations that run multiple models and need reproducibility; free tier for academic and personal use
Try it ↗SQL-based data transformation tool — standard in modern data stacks for building and documenting data models that feed analytics and ML; free dbt Core training helps data scientists understand the analytics engineering layer their models depend on
Try it ↗Collaborative data notebook with AI code assistance — combines SQL, Python, and no-code charts in one shareable workspace; AI features generate code from plain-language descriptions; used at data-forward organisations for analysis and stakeholder communication
Try it ↗AutoML platform that builds, evaluates, and explains predictive models from structured data — understanding how AutoML tools work and where they fall short is important context for data scientists who may be evaluated alongside them
Try it ↗Free online textbook on causal inference methods — difference-in-differences, instrumental variables, regression discontinuity; these methods for answering causal questions are what distinguish rigorous data science from correlation fishing
Try it ↗Extinction Timeline
GitHub Copilot and similar code assistants are handling the boilerplate coding work that used to consume significant data scientist time. The benefit is productivity — experienced data scientists can move faster. The risk is that organisations with simple modelling needs can now get results without hiring a dedicated scientist.
AutoML capabilities continue improving — the gap between what a skilled data scientist builds manually and what AutoML produces on standard supervised learning tasks is narrowing. The differentiator shifts further toward problem definition, experimental design, and business translation rather than technical model-building execution.
Data science as a standalone function is evolving. More organisations are building ML engineering (deployment-focused) and analytics engineering (data pipeline and transformation-focused) roles separately from the research scientist role. Data scientists who develop deep domain expertise in a specific industry vertical — healthcare, fintech, e-commerce operations — combined with strong business communication are in a more differentiated position than generalists.
AI is replacing the code-writing and standard model-building parts of the role efficiently. What it hasn't replaced is the judgment layer: deciding what to model, interpreting results in business context, designing rigorous experiments, and getting organisations to trust and act on model outputs. Data scientists who build those skills are in a substantially more durable position than those who focus only on model execution.
Causal inference and experimental design — understanding when correlation is and isn't causal, and how to design studies that answer business questions with appropriate confidence. Business communication — translating statistical findings into decisions that executives and product teams can act on. Domain expertise — deep understanding of a specific industry that makes your interpretation of results more valuable than a generalist. Statistical thinking that goes beyond picking the right sklearn function.
AutoML reduces the need for data scientists on standard supervised learning problems — classification, regression, forecasting on structured data with clear labels. It hasn't eliminated demand; it's shifted what organisations hire for. The roles being created are at the ends of the pipeline: data engineering (clean, labelled data for AutoML to use) and ML engineering (deployment, monitoring, governance of models in production). Exploratory research and novel problem formulation still require human expertise.
Understanding MLOps — how models are deployed, monitored, and retrained — is increasingly expected rather than optional. A model that exists only in a notebook isn't delivering business value. Data scientists who can take their work to production (or work closely with ML engineers to do so) are more valuable than those who hand off a Jupyter notebook. Tools: MLflow for experiment tracking, Docker basics, and familiarity with cloud ML platforms (AWS SageMaker, Azure ML, Vertex AI).
Take the free Fossil Score assessment at DontGoDinosaur.com. It looks at your specific daily tasks — not just your job title — and gives you a personalised risk score with practical steps for the next 6 months. It takes about 4 minutes.
More in Computer & Mathematical
Software Quality Assurance Analysts and Testers
AI helps software quality assurance analysts and testers do their jobs better and faster, but it can't replace the human skills at the heart of this work.
Mathematical Scientists
AI handles routine computation, literature search, and standard modelling. Mathematical scientists who do novel theoretical work or complex problem formulation are well positioned — those doing repetitive applied analysis face real pressure.
Database Administrators
Cloud-managed database services have automated a large part of routine DBA work — backups, patching, scaling, and performance tuning assistance are now platform features. DBAs who understand the platform deeply, manage complex environments, and handle security and architecture decisions are in a much more durable position than those doing only routine maintenance.
Mathematicians
AI helps mathematicians do their jobs better and faster, but it can't replace the human skills at the heart of this work.
Computer and Information Research Scientists
AI can write code and run experiments, but formulating genuinely novel research questions, designing studies, and advancing the field's understanding still require a trained researcher.
Aerospace Engineering and Operations Technologists and Technicians
AI is taking over the monitoring, documentation, and routine checklist work. The hands-on assembly, fault diagnosis, and safety judgment that aerospace standards demand still need a trained human in the room.
Further reading
Your Personal Score
Get a Fossil Score built on your actual daily tasks, not a category average. 4 minutes. Free.
Calculate My Personal Fossil Score