Data Science Solution Specialist – Generative AI

Deloitte

Data Scientist (Generative AI) -Solution Specialist – USDC

Are you an experienced, passionate pioneer in technology – a solutions builder, a roll-up-your-sleeves technologist who wants a daily collaborative environment, think-tank feel and share new ideas with your colleagues – without the extensive demands of travel? If so, consider an opportunity with our US Delivery Center – we are breaking the mold of a typical Delivery Center.

Our US Delivery Centers have been growing since 2014 with significant, continued growth on the horizon. Interested? Read more about our opportunity below …

Work youll do

The Generative AI Engineer will, as part of several client delivery teams, be responsible for developing, designing, and maintaining cutting-edge AI-based systems, ensuring smooth and engaging user experiences. Additionally, the Generative AI Engineer will participate in a wide variety of Natural Language Processing activities, including refining and optimizing prompts to improve the outcome of Large Language Models (LLMs), and code and design review. The kinds of activities performed by the Prompt Engineer will also include, but not be limited to:

Working across client teams to develop and architect Generative AI solutions using ML and GenAI

Developing and promoting standards across the community

Evaluating and selecting appropriate AI tools and machine learning models for tasks, as well as building and training working versions of those models using Python and other open-source technologies

Working with leadership and stakeholders to identify AI opportunities and promote strategy.

Developing and conducting trainings for users across the Government & Public Services landscape on principles used to develop models and how to interact with models to facilitate their business processes.

Building and prioritizing backlog for future machine-learning enabled features to support client business processes.

Youll design and build generative models, selecting the most suitable architecture (e.g., GANs, VAEs) based on the desired output (text, images, code). This involves writing code using Python libraries like TensorFlow or PyTorch.

Once your model is built, youll train it on the prepared data, fine-tuning hyperparameters to achieve optimal performance. Youll then evaluate the models outputs to assess its effectiveness and identify areas for improvement.

Youll collaborate with other engineers to integrate your generative AI solution into existing systems or develop new applications. This might involve deploying the model on cloud platforms for scalability.

The field of generative AI is rapidly evolving. Staying abreast of the latest research, advancements, and ethical considerations in AI development is an ongoing process.

The TeamArtificial Intelligence & Data Engineering

In this age of disruption, organizations need to navigate the future with confidence, embracing decision making with clear, data-driven choices that deliver enterprise value in a dynamic business environment.

The Artificial Intelligence & Data Engineering team leverages the power of data, analytics, robotics, science and cognitive technologies to uncover hidden relationships from vast troves of data, generate insights, and inform decision-making. Together with the Strategy practice, our Strategy & Analytics portfolio helps clients transform their business by architecting organizational intelligence programs and differentiated strategies to win in their chosen markets.

Artificial Intelligence & Data Engineering will work with our clients to:

Implement large-scale data ecosystems including data management, governance and the integration of structured and unstructured data to generate insights leveraging cloud-based platforms

Leverage automation, cognitive and science-based techniques to manage data, predict scenarios and prescribe actions

Drive operational efficiency by maintaining their data ecosystems, sourcing analytics expertise and providing As-a-Service offerings for continuous insights and improvements

Qualifications

Required:

3 years of experience programming in Python or R.

Knowledge of Python libraries like Pandas, Scikit-Learn, Numpy, NLTK is required

3 years of experience with Natural Language Processing (NLP) and Large Language Models (LLM) 3 years of experience building and maintaining scalable API solutions

Experience working with RAG technologies and LLM frameworks (Langchain, Claude and LLamaIndex), LLM model registries (Hugging Face), LLM APIs, embedding models, and vector databases (FAISS , Milvus , OpenSearch, Pinecone etc.)

Experience working with Retrieval Augmented Thoughts (RAT) and chain of thoughts.

Experience building scalable data models and performing complex relational databases queries using SQL (Oracle, MySQL, PostGres), etc.

Experience working with cloud computing platforms (e.g., AWS, Azure, Google Cloud) and containerization technologies (e.g., Docker, Kubernetes).

Utilize tools such as Docker, Kubernetes, and Git to build and manage AI pipelines

Experience driving DevOps and MLOps practices, covering continuous integration, deployment, and monitoring of AI

Experience with machine learning libraries and services like TensorFlow, PyTorch, or Amazon SageMaker.

Experience integrating GenAI solution on cloud platform (e.g., AWS, Azure, Google Cloud)

3 years of experience designing solutions to address client requirements

1 years of experience with the design and implementation (building, containerizing, and deploying end to end automated data and ML pipelines) of automated cloud solutions

3 years of experience in developing algorithms using data science technologies to build analytical models

3 years of data extraction/manipulation experience using scripts specific to AI/ML

3 years of modeling experience using a variety of regression and supervised and unsupervised learning techniques.

3 years of experience in data wrangling/cleansing, statistical modeling, and programming

3 years of extensive experience working in an Agile development environment

3 years of experience for fluency in both structured and unstructured data (SQL, NOSQL)

3 years of production experience with Apache Spark

3 years of hands-on experience with web APIs, CI/CD for ML, and Serverless Deployment

3 years of experience with presentation and data analysis software such as: SAS, R, SPSS, MATLAB, QlikView, Excel and Access

1 years of experience/familiarity with Linux OS and Windows servers

1 years of experience with/knowledge of Docker, Jenkins, Kubernetes, and other DevOps tools

Must live in a commutable distance (approximately 100-mile radius) to one of the following Delivery locations: Atlanta, GA; Charlotte, NC; Dallas, TX; Gilbert, AZ; Houston, TX; Lake Mary, FL; Mechanicsburg, PA; Philadelphia, PA; with the ability to commute to assigned location for the day, without the need for overnight accommodations

Expectation to co-locate in your designated Delivery location up to 30% of the time based on business needs. This may include a maximum of 10% overnight client/project travel

Ability to travel 10%, on average, based on the work you do and the clients and industries/sectors you serve. This may include overnight travel

Bachelors degree, preferably in Computer Sciences, Information Technology, Computer Engineering, or related IT discipline; or equivalent experience

Must be legally authorized to work in the United States without the need for employer sponsorship, now or at any time in the future

Preferred:

Previous Government Consulting and/or professional services experience

In depth understanding of AI protocols and standards

Understanding of technology risks and the ability to assess and mitigate them

Deep knowledge of a specific domain or industry, with a focus on applying NLP/LLM solutions in that context

Experience with debugging and troubleshooting software or solutions design issues

Proven ability to stay current with best practices and new technology solutions in the field

Ability to display both breadth and depth of knowledge regarding functional and technical issues

Experience presenting to clients or other decision makers to present and sell ideas to various audiences (technical and non-technical)

Certification from any of the three major cloud platforms (AWS / Azure / GCP) in Cloud Architecture / Engineering / DevOps / ML.

Familiarity with Kubeflow or MLflow

Experience with machine learning pipelines (Azure ML)

Familiarity with the latest Natural Language Processing or Computer Vision related algorithms

Information for applicants with a need for accommodation: https://www2.deloitte.com/us/en/pages/careers/articles/join-deloitte-assistance-for-disabled-applicants.html

All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability or protected veteran status, or any other legally protected basis, in accordance with applicable law.

Show Full Vacancy