Find
Try It
Search this page
Next
Future Job Report
Top 5 Technical Skills Related to AI
- Machine Learning (ML) Engineering: Designing, training, and deploying ML models using frameworks like TensorFlow or PyTorch.
- Natural Language Processing (NLP): Developing systems that understand and generate human language, such as chatbots or translation tools.
- Big Data Analytics: Handling and analyzing large datasets using tools like Apache Spark, Hadoop, or SQL to extract insights for AI systems.
- Neural Network Architecture Design: Building and optimizing deep learning models, including convolutional and transformer-based networks.
- Cybersecurity for AI Systems: Securing AI models and data pipelines from adversarial attacks and breaches.
- Machine Learning Engineering: Over 2 years up to and including 4 years (SVP Level 7)
- Natural Language Processing (NLP): Over 2 years up to and including 4 years (SVP Level 7)
- Big Data Analytics: Over 1 year up to and including 2 years (SVP Level 6)
- Neural Network Architecture Design: Over 2 years up to and including 4 years (SVP Level 7)
- Cybersecurity for AI Systems: Over 1 year up to and including 2 years (SVP Level 6)
Note: These estimates reflect the time typically required to achieve average performance in a job setting, including formal education, training, and essential experience. They do not include orientation time for adapting to a specific workplace.
AI Challenges Tackled by Leading Tech Companies
🔵 Microsoft
- Customizing large language models (LLMs) for enterprise use
- Efficient adaptation using fine-tuning and RLHF (Reinforcement Learning from Human Feedback)
- Scaling AI systems for internal and external products
- Deep collaboration with OpenAI and integration into GitHub Copilot, Office, and Azure
🔴 Google
- Using AI for scientific discovery (e.g., genomics, quantum computing)
- Developing AI to optimize data center efficiency and sustainability
- Building foundational models like Gemini for multimodal reasoning
- Balancing open research with responsible deployment
🟣 Meta
- Pursuing AGI through frontier models and infrastructure scale
- Developing open-source tools like LLaMA and Segment Anything
- Struggling with product-market fit and transparency in AI
- Shifting from open research to more proprietary approaches
🟢 NVIDIA
- Accelerating AI workloads through hardware-software co-design
- Optimizing GPU memory and latency for large model inference
- Benchmarking LLMs on CUDA code generation and reasoning tasks
- Enabling AI infrastructure for the entire ecosystem
🟠 Amazon
- Personalizing shopping and media experiences with AI
- Optimizing fulfillment and logistics using robotics and ML
- Scaling foundation models (e.g., Amazon Titan, Alexa LLM)
- Democratizing AI access through AWS services
⚖️ Shared Challenges
- Scaling compute and infrastructure efficiently
- Ensuring fairness, transparency, and ethical AI use
- Bridging the gap between research and real-world deployment
- Navigating global regulations and public trust
⚡ Hard Problems AI Is Solving in the Electric Utility Sector
🔌 1. Grid Reliability and Resilience
- AI helps detect and respond to grid instabilities in real time.
- Predictive analytics identify potential failures before they cause blackouts.
- Machine learning models optimize load balancing across distributed energy resources.
📈 2. Demand Forecasting and Load Management
- AI improves short- and long-term electricity demand forecasting using weather, usage, and behavioral data.
- Helps utilities avoid overproduction or shortages, reducing operational costs and emissions.
🌞 3. Renewable Energy Integration
- AI manages the variability of solar and wind power by predicting generation patterns.
- Supports dynamic grid reconfiguration to accommodate distributed energy sources.
🛠️ 4. Predictive Maintenance
- AI analyzes sensor data from transformers, substations, and lines to detect wear and tear.
- Reduces unplanned outages and extends asset life by enabling condition-based maintenance.
🧠 5. Intelligent Grid Automation
- AI enables self-healing grids that automatically isolate faults and reroute power.
- Supports autonomous decision-making in grid operations and restoration.
🔐 6. Cybersecurity and Risk Management
- AI detects anomalies in network traffic and operational data to prevent cyberattacks.
- Helps secure critical infrastructure from emerging threats as digitalization increases.
🏭 7. Managing AI’s Own Energy Demand
- Ironically, AI data centers are becoming major electricity consumers.
- Utilities must forecast and supply power to hyperscale AI infrastructure while maintaining grid stability.
🧩 8. Regulatory and Ethical Complexity
- AI must operate within strict regulatory frameworks for safety, transparency, and fairness.
- Utilities face challenges in deploying AI responsibly while ensuring public trust.
🎓 Hard Problems AI Is Expected to Solve in Education
📚 1. Personalized Learning at Scale
- AI can tailor content, pacing, and feedback to individual student needs.
- Helps address diverse learning styles, speeds, and knowledge gaps.
- Challenge: Avoiding bias and ensuring personalization doesn’t reinforce inequality.
🌍 2. Equity and Access
- AI can expand access to quality education in underserved regions.
- Translates content across languages and adapts to different cultural contexts.
- Challenge: One-third of the world remains offline, and AI tools often favor dominant languages and cultures.
🧠 3. Intelligent Tutoring and Feedback
- AI tutors can provide instant, adaptive feedback in subjects like math, science, and writing.
- Supports students outside classroom hours and reduces teacher workload.
- Challenge: AI still struggles with nuance, creativity, and emotional intelligence.
📝 4. Assessment and Grading Reform
- AI can automate grading and detect patterns in student performance.
- Enables formative assessment and real-time intervention.
- Challenge: Risk of over-reliance on standardized metrics and lack of transparency in scoring.
🧩 5. Curriculum Design and Content Generation
- AI can generate lesson plans, quizzes, and learning materials on demand.
- Supports differentiated instruction and teacher creativity.
- Challenge: Ensuring content accuracy, coherence, and alignment with learning goals.
🔐 6. Ethics, Privacy, and Data Governance
- AI systems rely on sensitive student data to function effectively.
- Challenge: Protecting privacy, ensuring consent, and preventing surveillance or misuse of data.
🌱 7. Sustainability and Infrastructure
- AI can optimize energy use in schools and support climate education.
- Challenge: Training large models consumes massive energy—raising environmental concerns.
The Hard Problem Hitachi Energy Is Hiring to Solve
Build a unified, high‑trust data backbone for the electric grid so
real‑time markets, forecasting engines, and control systems can make
fast, reliable decisions at scale.
What makes it hard
• Fragmented critical data: Market bids, grid models, telemetry, and forecasts live in separate, legacy systems.
• Real‑time pressure: Decisions must be correct within seconds, not minutes or hours.
• Heavy analytics stack: High‑performance engines (FORTRAN/C++/Python) need clean, consistent inputs to run optimization and simulation at scale.
• Reliability over hype: Any AI or automation must sit on top of a data layer that never breaks grid stability or market fairness.
• Modernizing the old: Decades of existing tools and workflows must be made “AI‑ready” without disrupting operations.
In one line
Turn chaotic grid and market data into a single, fast, trustworthy substrate
that future AI and automation can safely depend on.
From Grid Data Backbone → Human Recall Backbone
Hitachi Energy is tackling one of the hardest problems in the electric grid:
building a unified, high‑accuracy, high‑speed data layer that real‑time markets,
forecasting engines, and grid‑stability systems can trust.
INV‑BAT‑AI mirrors this challenge in the human domain.
Where utilities struggle with fragmented, high‑velocity operational data,
learners struggle with fragmented, high‑volume knowledge.
Both require a deterministic backbone that never fails under pressure.
How This Fits the INV‑BAT‑AI Strategic Framework
• Backbone Principle:
The grid needs a trusted data substrate; humans need a trusted recall substrate.
INV‑BAT‑AI becomes the “memory grid” for every learner and worker.
• Deterministic Reliability:
Grid operators cannot tolerate uncertainty in data.
Students and professionals cannot tolerate uncertainty in recall.
INV‑BAT‑AI provides exam‑grade and job‑grade reliability.
• High‑Velocity → High‑Volume Parallel:
Grid data streams are fast; human learning streams are massive.
Both require compression, organization, and instant retrieval.
• Automation Readiness:
The grid’s data layer enables AI‑driven forecasting, optimization, and self‑healing.
INV‑BAT‑AI’s recall layer enables AI‑enhanced cognition, faster problem‑solving,
and higher‑order thinking.
• Strategic Moat:
Whoever owns the trusted backbone—data for machines or recall for humans—owns
the next layer of intelligence.
INV‑BAT‑AI positions itself as the recall infrastructure for the future workforce.
The Hard Problem GE Vernova Is Hiring to Solve
Build a unified AI backbone that aligns data, teams, and workflows across
GE Vernova’s global energy businesses so AI can reliably improve Safety,
Quality, Delivery, and Cost at enterprise scale.
What makes it hard
• Fragmented operations: Power, wind, grid, and electrification all use different systems and data.
• Enterprise‑wide AI: Models must work across multiple business units with different constraints.
• Verification & risk: AI must be measurable, safe, and aligned with SQDC outcomes.
• Cross‑functional orchestration: Data scientists, IT, designers, and business leaders must move as one.
• 3rd‑party integration: External AI tools must be evaluated, aligned, and absorbed into the platform.
In one line
Build the AI operating system that powers GE Vernova’s energy transition.
The Hard Problem GE Vernova Is Hiring to Solve
Build advanced Distribution Management System (DMS) applications that keep
the modern grid stable as it becomes more dynamic, DER‑heavy, and dependent
on real‑time automation.
What makes it hard
• Dynamic distribution networks: High DER penetration makes
power flow, voltage, and reliability harder to control in real time.
• Mission‑critical applications: Functions like FLISR,
Volt‑VAR optimization, fault location, and feeder reconfiguration must be
fast, correct, and deterministic.
• Complex system integration: DMS, SCADA, DERMS, and
modeling tools must interoperate cleanly across utilities, ISOs, and
operators.
• High‑stakes software engineering: Grid‑operations software
must be designed, tested, tuned, and delivered with zero tolerance for
instability.
• Cross‑functional technical leadership: The role must guide
system engineers, frontend developers, and application engineers through
complex design decisions.
In one line
Deliver the next generation of DMS applications that keep the distribution
grid reliable as it transforms faster than ever.
The Hard Problem GE Vernova Is Hiring to Solve
Help utilities transition from siloed, legacy OT/IT systems to a unified,
cloud‑ready, data‑centric GridOS architecture that supports modern,
interoperable, real‑time grid operations.
What makes it hard
• Legacy → Cloud-native shift: Utilities must evolve from
traditional EMS/DMS/SCADA stacks to modular, API‑driven, microservices
architectures.
• Interoperability reality: GridOS must integrate with
everything from modern APIs to decades‑old file‑based exchanges.
• Data fabric adoption: Moving utilities from point‑to‑point
integrations to a shared, resilient data fabric is a major cultural and
technical leap.
• Cybersecurity by design: NERC CIP, Zero Trust, and
multi‑zone architectures must be embedded into every deployment.
• Enterprise-scale reliability: Solutions must work across
hybrid, multi‑site, and cloud environments with DevOps/DataOps resilience.
• Cross-functional orchestration: Architects must align
product, engineering, cybersecurity, commercial, and pre‑sales teams.
In one line
Architect the modern grid platform—secure, interoperable, cloud-native,
and ready for the energy transition.
The Hard Problem Black & Veatch Is Hiring to Solve
Transition legacy, hardware‑bound substations into secure, interoperable,
virtualized IEC‑61850 digital substations that deliver deterministic,
real‑time performance at production scale.
What makes it hard
• Virtualization leap: Protection and automation must run on
software‑defined platforms without losing millisecond‑level reliability.
• Interoperability reality: Mixed fleets of IEDs must behave
consistently under IEC 61850 models and messaging.
• Deterministic networking: PRP/HSR, VLANs, QoS, multicast control,
and PTP timing must be engineered with precision.
• Cyber‑informed design: Security must be embedded into the
architecture from the first diagram, not added later.
• Industry alignment: Utilities, OEMs, standards bodies, and the
vPAC Alliance must converge on shared digital‑substation practices.
In one line
Build the production‑ready digital substation — virtualized, secure,
interoperable, and ready for the future grid.
The Hard Problem AWS Is Hiring to Solve
Ensure AWS can energize massive data‑center loads on schedule across every
ISO and RTO in the Americas while grid codes, interconnection rules, and
market structures evolve faster than infrastructure can be built.
What makes it hard
• Rapidly changing grid codes: New NERC standards, PJM’s
expedited tracks, ERCOT’s Batch Zero, and state PUC reforms shift the
rules mid‑project.
• Hyperscale load integration: AWS data centers behave like
large industrial loads requiring deep compliance with VRT, FRT, SSO,
reactive power, and power‑quality requirements.
• Congested interconnection queues: AWS must energize on
time despite multi‑year queue delays and shifting study requirements.
• Market‑driven risk: Capacity markets, resource adequacy
proposals, and reliability procurement rules directly affect AWS’s cost
and timing.
• Regulatory influence: AWS must actively shape policy in
PJM, ERCOT, FERC, and state dockets to protect energization timelines.
• Mission‑critical reliability: Healthcare, emergency
services, finance, and global cloud operations cannot lose power — ever.
In one line
Secure reliable, compliant, on‑time grid interconnections for hyperscale
data centers in a rapidly changing regulatory landscape.
The Hard Problem Anthropic Is Hiring to Solve
Push compute utilization of Anthropic’s TPU/GPU fleet to the edge of the
physical envelope by unifying power engineering, cooling systems, workload
scheduling, and real‑time telemetry into one coordinated control layer.
What makes it hard
• Extreme AI loads: Accelerator clusters create massive,
rapidly changing power and thermal demands.
• Operating near physical limits: Utilization must be pushed
as high as safely possible without violating availability commitments.
• IT + OT convergence: Power distribution, cooling, telemetry,
and workload schedulers must operate as one system.
• Real‑time telemetry: SCADA/BMS/EPMS data must feed models
and control loops with millisecond‑level responsiveness.
• Reliability is absolute: Claude training runs cannot fail;
uptime is a first‑order requirement.
• Advanced modeling: Forecasting consumption, failure modes,
and oversubscription risk requires statistical modeling and simulation.
• Partner ecosystem: Data‑center providers must be pushed to
redesign architectures for AI‑era density and performance.
In one line
Engineer the control systems that turn raw data‑center capacity into
maximally efficient, highly reliable AI compute.
The Hard Problem Google Is Hiring to Solve
Engineer next‑generation high‑voltage and medium‑voltage electrical
infrastructure — substations, switchgear, transformers, microgrids — that
can safely and efficiently power hyperscale data centers with
mission‑critical reliability.
What makes it hard
• Hyperscale power demand: Google facilities require
utility‑grade HV/MV engineering to support massive continuous loads.
• Advanced substation design: AIS/GIS substations, grounding
systems, CT sizing, and detailed power‑system studies must be engineered
precisely.
• Hybrid AC/DC architectures: Google is pushing into
large‑scale DC distribution and microgrid‑ready designs.
• New electrical products: Switchgear, transformers,
energy‑storage systems, and microgrids must be developed for
mission‑critical environments.
• Zero‑downtime reliability: Global services depend on
uninterrupted power — failure is not an option.
• Cross‑discipline integration: Electrical, mechanical,
controls, civil, and IT/telecom systems must operate as one coherent
infrastructure.
• R&D + deployment: Designs must be innovative enough for
Google’s R&D lab yet robust enough for global rollout.
In one line
Build the utility‑grade electrical backbone that powers Google’s global
data‑center fleet.
The Hard Problem Google Is Hiring to Solve
Build the AI‑powered operations layer for Google Distributed Cloud so
operators can deploy, monitor, troubleshoot, and scale edge and on‑prem
cloud systems with automation, intelligence, and mission‑critical
reliability.
What makes it hard
• Distributed environments: GDC runs in customer data centers,
partner facilities, and edge sites — each with unique constraints.
• AI‑driven operations: AI must meaningfully improve deployment,
monitoring, troubleshooting, and lifecycle management for operators.
• Public‑sector demands: Security, compliance, and reliability
expectations are extremely high.
• Cross‑functional orchestration: PMs must align engineering,
design, support, marketing, and customers around one roadmap.
• Risk + bottleneck management: The role requires identifying
technical risks, scaling limits, and infrastructure bottlenecks.
• AI infrastructure expertise: GPUs, virtualization,
containerization, and agentic AI understanding are essential.
• Zero‑downtime expectations: GDC supports mission‑critical
workloads — outages are unacceptable.
In one line
Build the AI brain that powers Google’s distributed, hybrid, and edge cloud.
The Hard Problem Microsoft Is Hiring to Solve
Build the trust, safety, and reliability layer for Microsoft’s entire AI
ecosystem — ensuring every model, tool, and developer workflow can identify,
measure, mitigate, and monitor AI risks at planetary scale.
What makes it hard
• Planet‑scale AI: Safety must work across GitHub, VS Code,
Azure, Copilot, and enterprise workloads.
• Multimodal risk surface: Text, image, audio, video, and
multimodal models each introduce unique failure modes.
• Rapid iteration: The role demands constant prototyping,
experimentation, and shipping in fast cycles.
• Developer‑facing tooling: Safety must be embedded directly
into the tools developers already use.
• Cross‑product orchestration: Features must land across
multiple teams, orgs, and product lines.
• High availability: Safety services must be as reliable as
the AI systems they protect.
• Evolving threat landscape: New model capabilities create
new risks — the safety layer must adapt continuously.
In one line
Build the Responsible AI backbone that keeps Microsoft’s AI ecosystem safe
at global scale.
The Hard Problem Microsoft Is Hiring to Solve
Build the multimodal safety layer for Microsoft’s frontier‑scale AI models —
ensuring image, video, audio, and text systems behave safely when served to
millions of Copilot users every day.
What makes it hard
• Frontier multimodality: Safety must work across diffusion,
image, video, audio, and LLM models simultaneously.
• Post‑training risk discovery: The role requires uncovering
hidden failure modes that only appear after large‑scale training.
• Evaluation frameworks: Microsoft needs new red‑teaming,
stress‑testing, and robustness frameworks for multimodal systems.
• Safety‑focused fine‑tuning: Engineers must design
fine‑tuning and alignment algorithms specifically for multimodal safety.
• Automated guardrails: Safety pipelines must be reusable,
automated, and production‑ready for Copilot‑scale deployment.
• User‑validated safety: Safety decisions must be grounded
in real user needs and validated through research.
• Fast‑paced applied research: The team operates on the
bleeding edge — prototyping, testing, and shipping rapidly.
In one line
Build the multimodal safety backbone that keeps Microsoft’s next‑generation
AI models trustworthy at global scale.
March 2026
Labor market impacts of AI: A new measure and early evidence
Source: Anthropic
|
January 2026
How Automation and Augmentation Are Combining to Reshape Work
Source: The Burning Glass Institute
|
February 2026
How to build pro-worker AI
Source: MIT Sloan School of Management
|
April 2025
Skills and Workforce Development - April 2025
Source: World Bank Group
|
January 2025
Future of Jobs Report 2025
Source: McKinsey Digital
|
June 2023
Future of Jobs Report 2023
Source: McKinsey Digital
|
January 2025
Future of Jobs Report 2025
Source: World Economic Forum
|
May 2023
Future of Jobs Report 2023
Source: World Economic Forum
|
2025 Global Emerging Skills
Source: World Economic Forum
|
| Skills Proficiency
Skills Proficiency Level = Foundational, Experienced, and Advanced
|
| Machine Learning (Level 5)
Machine Learning Level 5 skill = Experienced and Advanced
|
| Analytical Thinking (Level 4)
Analytical Thinking Level 4 skill = Experienced and Advanced
|
| Creative Thinking (Level 4)
Creative Thinking Level 4 skill = Experienced and Advanced
|
| System Thinking (Level 4)
System Thinking Level 4 skill = Experienced and Advanced
|
| Artificial Intelligence (Level 4)
Artificial Intelligence Level 4 skill = Experienced and Advanced
|
| Computational Thinking (Level 4)
Computational Thinking Level 4 skill = Experienced and Advanced
|
| Computer Hardware and Networking (Level 4)
Computer Hardware and Networking Level 4 skill = Experienced and Advanced
|
| Problem-Solving (Level 3)
Problem-Solving Level 3 = Experienced, and Advanced
|
| Problem-Solving (Level 3)
Problem-Solving Level Career Advice
|
| Technology Design and Programming (Level 3)
Technology Design and Programming Level 3 = Experienced, and Advanced
|
| Innovation and Creativity (Level 2)
Innovation and Creativity Level 2 = Experienced, and Advanced
|
| July 2022
Intel – AI for Workforce Program
|
| July 2022
The Future Workforce
|
SKILLS SIMILARITY TO AI/ML SKILLS - May 2022
SKILLS SIMILARITY TO AI/ML SKILLS
Source:
Burning Glass Institute,
the Business-Higher Education Forum, and Wiley
|
SKILLS SIMILARITY TO CLOUD SKILLS - May 2022
SKILLS SIMILARITY TO CLOUD SKILLS
Source:
Burning Glass Institute,
the Business-Higher Education Forum, and Wiley
|
HIGH DEMAND CLOUD SKILLS - May 2022
HIGH DEMAND CLOUD SKILLS
Source:
Burning Glass Institute,
the Business-Higher Education Forum, and Wiley
|
| A “State of Skills” Report from the Burning Glass Institute 2022
A “State of Skills” Report from the Burning Glass Institute,
the Business-Higher Education Forum, and Wiley |
| These skills are in high demand - BHEF 2022
Among the bulwarks for all workers are the foundational skills—the ability to set
and achieve goals, manage projects, make sense of data, communicate effectively,
and work well with teams. These skills are in high demand, lead to higher pay, afford
workers greater mobility, and increase in value over time.
|
| Business Higher Education Forum (BHEF) 2022
Business Higher Education Forum (BHEF) members, including Wiley,
PwC, and Ellucian, are involved in customer discovery and pilot testing,
with upcoming opportunities for other members to engage
|
| May 2021
AI Related Curriculum to Help Meet Emerging Workforce
|
| January 2021
Upskilling for Shared Prosperity
|
| November 2020
What Microsoft’s Satya Nadella thinks about work of the future
|
|
| January 2020
Intel Digital Readiness Programs, AI for Youth, AI for Citizens, AI for Current Workforce, and AI for Future Workforce
|
| January 2020
I believe US companies should invest in the future AI workforce
|
| AI Readiness Index 2020 Report
AI Readiness Index 2020 Report
|
| July 2019
Building Workforce Skills for AI
|
| Digital Skills - 2019
Digital Skills Analysis Around The World
. |
| Deciphering China's AI Dream - 2018
Deciphering China's AI Dream
|
| August 2017
We’ll be part of a globally connected, remote workforce
|
| U.S. Bureau of Labor Statistics
2018 Standard Occupational Classification System (SOC) Definitions U.S. Bureau of Labor Statistics
|
| U.S. Bureau of Labor Statistics
2018 Standard Occupational Classification System (SOC). All workers are classified into one of 867 detailed occupations according to their occupational definition. To facilitate classification, detailed occupations are combined to form 459 broad occupations, 98 minor groups, and 23 major groups. Detailed occupations in the SOC with similar job duties, and in some cases skills, education, and/or training, are grouped together.
|
| Student and Exchange Visitor Information System (SEVIS)
Schools throughout the United States and its territories offer thousands of similar programs of study, but each have their unique program names and descriptions. The U.S. Department of Education (ED), which collects and reports data about these programs of study, developed the classification of instructional programs (CIP) ⤵ |
| U.S. Workforce Training Expenditures
Annual Expenditures for Workforce Education and Training in the U.S., 2017 Estimates |
| Knowledge, skills, and abilities
Knowledge, skills, and abilities
|
| Credential Engines
Credential Engines
|
| Credential Transparency Description Language (CTDL)
Colleges, job training programs, and employers offer all kinds of credentials—from
degrees to certifications to badges—for learning and gaining skills. But with almost
1,000,000 different credentials it is hard to easily find reliable and actionable
information on the value of various credentials, where to get them, what each
competency and skill means, cost, and employment opportunities associated with
each opportunity. Credential transparency requires bringing together this varied
information from colleges, companies, government agencies and others to describe
the knowledge and skills they confer and expect in clear, detailed, and consistent
language—the Credential Transparency Description Language (CTDL)
|
Start building your permanent memory system today.
Copyright 2026
Never Forget Again with IN-V-BAT-AI
INVenting Brain
Assistant Tools
using Artificial Intelligence
(IN-V-BAT-AI)
Since
April 27, 2009