Follow our Data scientist curriculum
and boost your career!
Eligible CPF and multi-financing up to 100%
To be recalled Access to the programmeApproach 3P
Our training centre guides you in identifying the ideal training, helping you maximize funding opportunities.
We put all the keys in hand for a start with confidence.
Experience an immersive and intensive training experience, designed to dive into practical workshops and real case studies.
Learn by doing, and develop concrete skills directly applicable to your future projects.
At the end of your career, we evaluate your acquired skills, issue certification attesting to your expertise, and accompany you to ensure your success in your professional projects.
You are now ready to excel!
Description of the training
intensive training to acquire the basic skills of data science, covering data collection and cleaning, visualisation, exploratory analysis, as well as machine learning and adapted tools.
Objectives of training
At the end of this training, participants will be able to:
- Block 1: Databases of Data Science - Understanding the basics of data science and the Python ecosystem. Apply statistical concepts and start in supervised machine learning.
- Block 2: Advanced Machine Learning Techniques - Mastering unsupervised algorithms and processing textual and temporal data. Optimize models and introduce Deep Learning.
- Block 3: Advanced projects and industrialization - Apply Deep Learning to computer vision and NLP. Deploy models and supervise them in production.
- Block 4: Final project and professional preparation - Design and implement a complete project. Validate skills and prepare for professional certifications.
Who is this training for?
The training is aimed at a wide audience, including:
The training is aimed at a wide audience, including:
- Developers and software engineers wishing to specialize in data science.
- Data analysts who want to strengthen their machine learning skills.
- Data engineers interested in managing and deploying data pipelines.
- Technical project leaders wishing to understand data science to better manage AI projects.
- Researchers and students in retraining or preparing professional certifications.
Prerequisites
No specific prerequisites are required.
Training programme
Block 1: Databases of Data Science (5 days)
- Objective: To understand the fundamentals of data science and the Python ecosystem.
- Content: Introduction to data science, main tools (Python, Jupyter), data manipulation with pandas.
- Objective: To explore data structures and types.
- Content: Exploratory analysis, missing data management, visualization with Matplotlib and Seaborn.
- Purpose: Apply mathematical bases to data analysis.
- Content: Descriptive statistics, hypothesis tests, distributions, introduction to probabilities.
- Purpose: Discover the techniques of supervised machine learning.
- Content: Linear regression, logistic regression, model evaluation (metric, overlearning).
- Objective: To master unsupervised algorithms.
- Content: Clustering (k-means, DBSCAN), reduced dimensionality (ACP, T-SNE).
- Objective: Manipulate time series and NLP-specific data.
- Content: Time series with pandas, introduction to NLP with spaCy and NLTK.
- Objective: Optimize model performance.
- Contents: General methods (random forests, gradient boosting), hyperparameter tuning with GridSearchCV.
- Objective: To understand the basics of neural networks.
- Content: Perceptron, backpropagation, setting up a model with TensorFlow.
- Purpose: Apply deep learning to computer vision.
- Content: Convolutive networks, applications to image classification.
- Objective: Use NLP tools.
- Content: Advanced techniques: Embeddings, transformations (BERT).
- Objective: Implement a deployment pipeline.
- Content: API with Flask, cloud integration (AWS Sagemaker, Azure ML).
- Objective: Supervise models in production.
- Content: Introduction to MLflow, data drift management.
- Purpose: Design an end-to-end project.
- Content: Problem definition, data exploration, installation of the processing pipeline.
- Objective: Implement an efficient solution.
- Content: Model training, tests and adjustments.
- Objective: Validate skills.
- Content: Project support, preparation for certifications (AWS, Azure, GCP).
Training assets
- Pedagogical and modular approach: Alternative between theory and practice for better assimilation of concepts.
- Cloud Integration: Strong focus on cloud and distributed solutions.
- Qualified speakers: Specialist trainers with practical experience in the field.
- Educational tools and materials: Access to online resources, live demonstrations and real-life case studies.
- Accessibility: Training is open to all, without advanced technical prerequisites.
- Implementation: Complete project from the end of the modules to consolidate the achievements.
- Preparation for Industry: Focus on standard certifications and tools used in the professional environment.
Pedagogical methods and tools used
- Live demonstrations with data science services.
- Practical workshops and real case studies in various sectors (industry, trade, health).
- Feedback: Sharing best practices and common mistakes in business.
- Simulations and tools: Using simulators for interactive workshops.
Evaluation
- End of training QCM to test the understanding of the concepts addressed.
- Practical case studies or group discussions to apply the knowledge gained.
- Ongoing evaluation during practical sessions.
- Implementation: Complete project from the end of the modules to consolidate the achievements.
Normative References
- Well-Architected Cloud Framework.
- GDPR (General Data Protection Regulation).
- ISO 27001, SOC 2 (Service Organization Control).
- NIST Cybersecurity Framework.