0% found this document useful (0 votes)
29 views6 pages

Elie Niring

Uploaded by

swatipmishra
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
Download as docx, pdf, or txt
0% found this document useful (0 votes)
29 views6 pages

Elie Niring

Uploaded by

swatipmishra
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1/ 6

Elie Niring

PROFESSIONAL SUMMARY
Innovative Generative AI Researcher & Comprehensive Machine Learning Developer with over 12 years
of expertise in leveraging deep learning, artificial intelligence, and statistical techniques to solve data
science challenges, enhancing organizational understanding, profitability, and market presence. Adept at
crafting algorithms, and deploying innovative solutions to complex business issues promptly, and
effectively; experienced in knowledge management systems, and language ontologies. Proficient in
executing solutions utilizing popular Generative AI, NLP frameworks, and libraries in Python (Langchain,
LlamaIndex, Hugging Face, NLTK, spaCy) or vector databases (Pinecone, FAISS). Familiar with the
implementation of Neural Networks, Support Vector Machines (SVM), and Random Forest techniques.
Keeps abreast of the latest advancements in data science, operations research, and Natural Language
Processing to ensure the use of cutting-edge techniques, algorithms, and technologies. Possesses
expertise in remote sensing; skilled in identifying, and creating suitable algorithms to uncover patterns,
and validate findings through experimental, and iterative methods. Strong interpersonal, and analytical
abilities, capable of multitasking, and adapting in high-pressure environments; a creative problem solver
with logical thinking skills, and keen attention to detail.

TECHNICAL EXPERIENCE
IDEs: Jupyter, Google Colab, PyCharm, R Studio

Programming: Python, R, SQL, MatLab

Python Libraries: Tensorflow, Pytorch, NLTK, Numpy, Pandas, OpenCV, Python Image
Library, Scikit-Learn, Scipy, Matplotlib, Seaborn, HuggingFace

NLP: Sentiment Analysis, Sentiment Classification, Sequence to Sequence


Model, Transformer, Bert, GPT 3.5

Analytical Methods: Exploratory Data Analysis, Statistical Analysis, Regression Analysis,


Time Series Analysis, Survival Analysis, Sentiment Analysis, Principal
Component Analysis, Decision Trees, Random Forest

Data Visualization: Matplotlib, Seaborn, Plotly, Folium

Computer Vision: Convolutional Neural Network (CNN), HourGlass CNN, RCNNs, YOLO,
Generative Adversarial Network (GAN)

Regression Models: Linear Regression, Logistic Regression, Gradient Boosting Regression,


L1 (Lasso), L2 (Ridge)

Tree Algorithms: Decision Tree, Bagging, Random Forest, AdaBoost, Gradient Boost,
XGBoost, Random Search and Grid Search

Cloud Data Systems: AWS, GCP, Azure

PROFESSIONAL EXPERIENCE
BAXTER – Deerfield, IL Sep 2023 – Present
Senior Data Scientist Architect – ML-Ops
In this project, I leveraged Elasticsearch for RAG implementation, enhancing few-shot prompting for the
Text2SQL engine. By deploying the application with FastAPI and Docker, I achieved reduced latency
through caching and automated scaling via Kubernetes. Utilizing advanced Large Language Models and
transformer architectures, analyzed consumer sentiment from platforms like Yammer and Cultura,
implementing classification algorithms to identify emerging trends. Additionally, I designed a CI/CD
pipeline to streamline data processing and model deployment, contributing to a solution that personalizes
Medicare plan recommendations during the annual enrollment period.
 Utilized Elasticsearch for RAG implementation, establishing it as a source for few-shot prompting to
the Text2SQL engine.
 Decreased latency in the analytics engine application by applying caching techniques.
 Deployed the application using FastAPI, and Docker containers, incorporating automated scaling via
the Kubernetes framework.
 Achieved a 30% reduction in LLM inference costs by utilizing Route LLM.
 Leveraged advanced Large Language Models (LLMs), and transformer-based architectures to
analyze patterns, and trends in consumer comments, and posts from platforms like Yammer, and
Cultura, aiming to capture real-time insights into sentiment, and emerging trends.
 Implemented classification algorithms to categorize consumer comments, and posts into predefined
topics, enabling the identification of trending content, and a deeper understanding of consumer
sentiment on specific subjects.
 Developed techniques to automatically generate new topics from Yammer, and Cultura posts, and
comments, enriching the topic classification system with more relevant, and dynamic categories.
 Seamlessly deployed, and managed both processes through Azure Pipelines, ensuring automated,
scalable, and efficient delivery.
 Played a pivotal role in a collaborative project, overseeing crucial stages such as developing an
unsupervised outlier identification algorithm, and implementing the CI/CD pipeline.
 Executed unsupervised outlier detection using a unique approach that involved five distinct outlier
detection methods to label the dataset, aggregating findings to create an outlier_percent column for
streamlined filtering.
 Developed robust, and scalable data science solutions in Python, focusing on data preprocessing,
feature engineering, and machine learning model development.
 Wrote efficient, and maintainable code for data analysis, model training, and deployment, adhering to
best practices in software development.
 Implemented, and optimized machine learning algorithms in Python, utilizing popular libraries such as
TensorFlow, PyTorch, Scikit-Learn, and Pandas.
 Leveraged Azure Cloud, Databricks, Jenkins, Docker, and Kubernetes to design, implement, and
manage a robust CI/CD pipeline for automated data ingestion, processing, model training, and
deployment.
 Developed, and deployed scalable machine learning models using Azure Machine Learning,
optimizing performance, and cost-effectiveness for tasks such as classification, regression, and
clustering.
 Hosted on Databricks, the pipeline seamlessly extracted raw data from Snowflake, conducted ETL
operations, identified outliers, and uploaded processed data to Azure Blob Storage.
 Designed, and deployed scalable machine learning solutions in Azure, utilizing services like Azure
Machine Learning, Azure Databricks, and Azure Synapse Analytics.
 Developed, and managed data pipelines using Azure Data Factory, and Azure Data Lake to support
data science workflows.
 Implemented, and optimized cloud-based infrastructure for data storage, processing, and model
deployment.
 Ensured the security, and compliance of data science applications by configuring, and managing
Azure Identity, and Access Management (IAM) roles, Key Vault, and encryption standards.
 Monitored, and optimized the performance of machine learning models in production using Azure
Monitor, and Application Insights.
 Leveraged transformer-based architectures like GPT, BERT, and T5 to build, and optimize models
capable of understanding, and generating human-like text.
 Designed, and implemented NLP pipelines that preprocess, tokenize, and clean large text corpora for
model training, and inference.
 Created, and managed complex SQL queries for extracting, transforming, and loading (ETL) large
datasets from relational databases, data warehouses, and cloud platforms.
 Optimized SQL queries, and database structures to enhance data retrieval, and analytics processes.
 Conducted data validation, cleansing, and preparation using SQL to ensure data quality, and integrity
for machine learning models.
 Orchestrated Jenkins tasks for data preparation, merging dataset files into a unified CSV, and storing
it in the workspace's datastore with versioning.
 Initiated model training tasks in Azure, executing code on a Databricks cluster, and saving the output
as a new model in AzureML.
 Successfully deployed models using an Embedded Architecture approach, integrating models with
the app within a Docker image for efficient deployment.
 Implemented creational design patterns in the CI/CD pipeline for reusability, and behavioral patterns
in algorithms, and integrations to enhance efficiency.
 Collaborated within a team structure led by a Data Science Manager, working alongside three Data
Scientists to achieve project objectives.
 Utilized tools such as Snowflake, Jenkins, Azure Cloud, Docker, Databricks, PySpark, and Twistlock
to streamline various aspects of the project.
 Adopted a canary deployment process, starting with limited access, and gradually expanding to
ensure smooth, and controlled model deployment.
 Contributed significantly to the overarching project goal of developing a workflow for personalizing
Medicare plan recommendations for members seeking new plans during annual enrollment.
 Leveraged Langchain, and Azure OPENAI to build a full-fledged scalable Generative AI application.
 Build a Natural language Data Analytics, and Text2SQL engine using Agentic workflow in langchain.
 Integrating RAG (Retrieval Augmented Generation) with the Natural language Data analytics, and
Text2SQL engine to respond to queries specific to the organization.

REGIONS FINANCIAL BANK – New York, NY Mar 2021 – Aug 2023


Senior AI Scientist
As a Senior AI Specialist at Regions Financial Bank, I spearheaded the development of innovative AI
solutions that enhanced customer engagement and operational efficiency. By deploying predictive
algorithms using Keras and TensorFlow, I facilitated a 9% increase in repeat customers through an
optimized recommendation system. I led cross-functional collaborations to ensure seamless integration of
machine learning models, while continuously monitoring performance for sustained accuracy. My role
also involved mentoring junior team members and conducting training sessions, fostering a culture of
knowledge sharing and skill development in AI technologies.
 Created a script for deploying updated Docker images to EC2 instances, ensuring timely, and efficient
updates.
 Led discussions with Regional Bank representatives to review project progress, explore concepts,
and resolve any blockers or issues.
 Utilized Keras, and TensorFlow to develop predictive algorithms, and address analytical challenges.
 Designed an NLP-based filter using embeddings, and BERT in TensorFlow, and Keras for advanced
text analysis.
 Analyzed large datasets to identify patterns and insights, informing the development of data-driven
solutions.
 Developed, and maintained machine learning models, and algorithms, ensuring high accuracy, and
reliability.
 Monitored model performance, and implemented continuous improvement processes for AI
applications.
 Provided technical guidance, and mentorship to junior data scientists, and AI specialists, fostering
skill development, and knowledge sharing.
 Conducted research on emerging AI technologies, and methodologies to drive innovation within the
organization.
 Conducted A/B testing that resulted in a 9% increase in repeat customers among users of the
recommendation system.
 Investigated new AI techniques, and methodologies, incorporating innovative solutions into existing
projects.
 Developed, and executed strategies for data collection, storage, and analysis, ensuring data quality,
and availability for machine learning applications.
 Oversaw the deployment of machine learning models into production environments, ensuring smooth
integration with existing systems.
 Continuously monitored, and tuned model performance, applying techniques such as hyperparameter
tuning, and feature selection to enhance predictive accuracy.
 Ensured that AI solutions adhered to ethical standards, and regulatory requirements, promoting
fairness, accountability, and transparency.
 Worked closely with business stakeholders to understand their needs, translating requirements into
effective AI solutions.
 Maintained thorough documentation of algorithms, processes, and code to facilitate knowledge
sharing, and support reproducibility in AI projects.
 Conducted training sessions, and workshops for staff to improve understanding, and capabilities in AI
technologies, and tools.
 Rapidly prototyped new AI solutions, and algorithms to validate concepts, and demonstrate their
feasibility to stakeholders.
 Defined, and tracked key performance indicators (KPIs) for AI initiatives, using data to measure
success, and guide decision-making.
 Collaborated with data engineers, software developers, and other teams to ensure cohesive
development, and deployment of AI solutions.
 Analyzed customer feedback, and behavior data to enhance AI models, tailoring solutions to meet
specific user needs.
 Assisted in budgeting for AI projects, estimating resource needs, and justifying expenditures based
on projected returns.
 Conducted risk assessments related to AI implementation, identifying potential challenges, and
developing mitigation strategies.
 Participated in AI conferences, seminars, and workshops to stay updated on industry trends, and
contribute to the AI community.
 Developed a Hybrid Mixed recommendation system by integrating collaborative filtering, content-
based, and demographic techniques.
 Executed A/B testing to identify the most effective recommendation system, addressing the "cold
start" challenge with a demographic-based approach.
 Employed Pandas, and NumPy for data preprocessing, cleaning, and feature engineering, utilizing
Python to manage missing values in the dataset.
 Applied text preprocessing techniques, such as stemming, and lemmatization, to optimize the corpus
for analysis.
 Implemented cloud-based data warehousing, and analytics solutions on GCP, leveraging tools like
Cloud Storage, Cloud SQL, and Looker for insights.

PUBLIX – Lakeland, FL Jan 2019 – Feb 2021


ML-Ops Engineer
As an ML-Ops Engineer at Publix, I developed a customized in-session product recommendation engine
to elevate the user experience. I automated text summarization and clustering tasks with Python and
contributed to Next-Best Offer predictions, designing Microassortments for future store layouts. My role
involved conducting anomaly detection and root cause analysis, merging consumer profiles for deeper
insights, and implementing end-to-end solutions for both batch and real-time algorithms, ensuring efficient
deployment and monitoring using tools like TensorFlow and Docker.
 Developed a tailored in-session product recommendation engine to enhance user experience.
 Created Python scripts to automate text summarization and clustering tasks.
 Contributed to Next-Best Offer predictions, and designed Microassortments for next-generation
stores.
 Conducted anomaly detection, and root cause analysis to identify, and resolve issues.
 Prepared data for integration with machine learning models, ensuring quality, and relevance.
 Merged consumer profiles using probabilistic record linkage to create comprehensive customer
insights.
 Led visual search initiatives to identify similar, and complementary products effectively.
 Designed, built, maintained, and optimized a comprehensive suite of algorithms, and their underlying
systems.
 Analyzed large datasets, applied machine learning techniques, and developed predictive, and
statistical models, utilizing best-in-class methodologies for enhancements.
 Implemented end-to-end solutions for both batch, and real-time algorithms, including monitoring,
logging, automated testing, performance testing, and A/B testing.
 Collaborated closely with data scientists, and analysts to roll out new product features on the e-
commerce website, in-store portals, and mobile applications.
 Established scalable, and efficient automated processes for data analysis, model development,
validation, and implementation.
 Deployed solutions using TensorFlow, Keras, Docker, and Elastic Kubernetes Service for seamless
integration.
 Executed strategies for monitoring model drift, and implemented retraining procedures.

DODGE – Auburn Hills, MI Aug 2017 – Dec 2018


Machine Learning Engineer
Improved operational efficiency and product quality through data-driven solutions. Developed and
deployed predictive models to optimize maintenance schedules, streamline supply chain logistics, and
automate quality control processes. Leveraged advanced machine learning techniques to enhance
vehicle performance and safety. Utilized data analytics to gain valuable insights into customer
preferences and market trends, informing product development and marketing strategies.
 Developed machine learning models to detect anomalies in vehicle sensor data, identifying potential
mechanical issues, and reducing maintenance costs.
 Implemented predictive maintenance models to predict equipment failures, and schedule preventive
maintenance, optimizing vehicle uptime, and reducing downtime costs.
 Utilized machine learning to optimize supply chain logistics, improve inventory management, reduce
lead times, and minimize transportation costs.
 Developed computer vision models to automate quality control inspections, identifying defects, and
inconsistencies in vehicle assembly.
 Analyzed customer reviews, and social media data to identify trends, preferences, and potential
issues, informing product development, and marketing strategies.
 Utilized Pandas to meticulously clean, and transform data, ensuring data quality, and accuracy for
subsequent analysis, and modeling.
 Extracted key features from vehicle sensor data, historical maintenance records, and other relevant
sources, identifying variables that contribute to predictive accuracy.
 Implemented advanced machine learning algorithms, including those from Scikit-learn, and
TensorFlow, to build predictive models for various automotive applications.
 Fine-tuned model performance through hyperparameter tuning using Scikit-learn, and AWS
SageMaker, maximizing accuracy, and efficiency.
 Seamlessly integrated machine learning models into existing systems, enabling real-time decision-
making, and automated processes.
 Improved model transparency through SHAP, and other interpretability techniques, enabling a better
understanding of model decisions, and identifying potential biases.
 Monitored key performance metrics, such as accuracy, precision, recall, and F1-score, to assess
model performance, and identify areas for improvement.
 Established a feedback loop to continuously retrain, and update models, ensuring their effectiveness
over time.
 Created comprehensive documentation, and collaborated with cross-functional teams to facilitate
knowledge transfer, and effective implementation of machine learning solutions.

NEW YORK LIFE INSURANCE – New York City, NY Nov 2015 – Jul 2017
Data Scientist
As a Data Scientist at New York Life Insurance, I developed customized product recommendations using
advanced machine learning algorithms, focusing on Collaborative Filtering to enhance customer
engagement and attract new clients. I led the design and deployment of various machine learning models,
including logistic regression and neural networks, while innovating optimization algorithms for diverse
applications. By conducting in-depth research on statistical techniques and leveraging tools like R and
Tableau for data visualization, I gained valuable insights into customer behavior, ensuring high data
integrity through meticulous cleaning and analysis.
 Developed tailored product recommendations by implementing sophisticated machine learning
algorithms, focusing on Collaborative Filtering to meet the unique needs of current customers, and
attract new ones.
 Led the design, and deployment of a variety of machine learning algorithms, utilizing techniques such
as logistic regression, random forest, KNN, SVM, neural networks, linear regression, lasso
regression, and k-means for comprehensive modeling.
 Innovated optimization algorithms specifically designed for data-driven models, broadening their use
across multiple machine learning approaches, including supervised, unsupervised, and reinforcement
learning.
 Conducted thorough research on statistical machine learning techniques, covering forecasting,
supervised learning, classification, and Bayesian methods, to integrate cutting-edge methods into the
modeling framework.
 Enhanced the technical complexity of solutions by incorporating machine learning, and advanced
technologies, leading to improved overall model performance.
 Performed exploratory data analysis, and created impactful data visualizations using R, and Tableau
to deepen insights into underlying data patterns.
 Collaborated effectively with data engineers to implement the ETL process, playing a key role in
optimizing SQL queries for efficient data extraction, and merging from Oracle databases.
 Utilized a diverse skill set in R, Python, and Spark to develop a range of models, and algorithms,
addressing various analytical needs within the project.
 Maintained data integrity through thorough checks, effective data cleaning, exploratory analysis, and
feature engineering, using both R, and Python to uphold high data quality standards.

CLOUDERA – Santa Clara, CA Jul 2012 – Oct 2015


Data Scientist
As a Data Scientist with Cloudera, I spearheaded the development of a vectorization function to capture
and embed facial features, significantly enhancing the representation of key facial attributes. I designed a
custom algorithm for efficient storage and comparison of these features, optimizing the verification
workflow. Improved image analysis depth while implementing robust data cleaning and augmentation
techniques. Finally, I ensured seamless deployment of the models through Flask and Pickle, making them
easily accessible for practical use.
 Developed a vectorization function to capture, and embed facial features, improving the
representation of essential facial attributes.
 Created a custom algorithm for the efficient storage, and comparison of vectorized features,
optimizing the verification workflow.
 Implemented image analysis techniques using traditional machine learning algorithms like Support
Vector Machines (SVMs), and Random Forests.
 Performed thorough data cleaning on both image, and tabular datasets to ensure high data quality
and accuracy.
 Employed image augmentation methods to introduce rotational, motion, and scale invariance for
more resilient model training.
 Established statistical evaluation methods to assess, and validate the performance of the models
created.
 Managed deployment through Flask, and Pickle, ensuring smooth integration, and accessibility of the
models.

EDUCATION
Master of Science in Information Technology
Carnegie Mellon University – 2012

You might also like