Senior Data Scientist

0 years

0.0 Lacs P.A.

Ahmedabad, Gujarat, India

Posted:1 week ago| Platform: Linkedin logo

Apply Now

Skills Required

datamarketinganalyticslearningmodeltestingoptimizationdesignengineeringlogicalgorithmseffectivenessexportvisualizationdocumentationarchitecturecodingcodeprocessinganalysisstoragecorrelationdevelopmentxgboosttuningevaluationstrategiesmetricsscoredeploymentdriveaigitcollaborativeloggingmonitoringreliabilityefficiencyscalabilitypythontensorflowsqlnosqlvertexpowertableau

Work Mode

On-site

Job Type

Full Time

Job Description

Company Description Tatvic, is a marketing analytics company focusing on generating insights from data using long association with Google & its infrastructure. We breed,recognize and reward performance. As a company we are growing very fast & we are under transformation. To enable this transformation we need future leaders with eyesight which balances execution & strategic understanding. Website: www.tatvic.com Mission The Senior Data Scientist will provide value to clients by deriving actionable insights from data through feature definition, machine learning model creation, testing and validation, optimization, and presenting results in an actionable format. The goal is to enable data-driven business decisions for clients. Role Responsibilities Responsibilities w.r.t Customer: Communicating with Customers to discover and understand the problem statement. Design a Solution that clearly aligns the Problem Statement, Solution Details, Output, Success Criteria and how the impact of the Solution aligns to a specific Business Objective. Designing a solution. Bird’s eye view of the Platform to create an effective solution. Feature Engineering: Identify features that would matter and ensure that the logic for feature selection is transparent & explainable. Model Selection: Use Pre-trained Models, AutoML, APIs or individual algorithms & libraries for effective model selection for optimal implementation. Optimize the model to increase its effectiveness with proper data cleansing and feature engineering refinements. Deploy the model for Batch or real-time predictions using methodologies like MLOps. Display or export the output into a visualization platform. Create a POC for providing data insight for the customer at short notice. Maintaining and Managing Project execution trackers and documentation. Keep the promises made to the customer in terms of deliverables, deadlines and quality. Innovation And Asset Building Responsibilities Design and Build reusable solutions that can be reused for multiple customers. Create clear documentation on architecture and design concepts & technical decisions in the project. Conduct internal sessions to educate cross-team stakeholders to improve literacy of the domain & solutions. Maintain coding standards & build reusable code & libraries for future use and enhancing Engineering at Tatvic. Stay up-to-date with innovations in data science and its applications in Tatvic relevant domains. Frequently Perform POCs to get hands-on experience with new technologies, including Google Cloud tools designed for Data Science applications. Explore the usage of data science in various business and web analytics applications Technical Skills Data Handling: Manage data from diverse sources, including structured tables, unstructured text, images, videos, and streaming/real-time data. For scalable data processing and analysis, utilize cloud platforms (preferably Google Cloud) such as BigQuery, VertexAI, and Cloud Storage. Feature Engineering: Identify and select relevant features with transparent and explainable logic. Design new derived features to enhance model performance and enable deeper insights. Utilize advanced techniques like Pearson Coefficient and SHAP values for feature importance and correlation analysis. Model Development: Select and build models based on problem requirements, using pre-trained models, AutoML, or custom algorithms. Experience in Linear, Non-linear, Timeseries (RNNs, LSTMs), Tree-based models (XGBoost, LightGBM), and other foundational approaches. Apply advanced modeling techniques like CNNs for image processing, RCNNs, YOLO for object detection, and RAGs and LLM tuning for text and search-related tasks. Optimize models with hyperparameter tuning, Bayesian optimization, and appropriate evaluation strategies. Model Evaluation: Assess model performance using metrics suited for data type and problem type: For categorical data: Precision, Recall, F1 Score, ROC-AUC, and Precision-Recall curves. For numerical data: Metrics like Mean Absolute Error (MAE), Root Mean Square Error (RMSE), Mean Squared Error (MSE), and R-squared (R²). Deployment & MLOps: Deploy models for batch or real-time predictions using MLOps frameworks, leveraging tools such as VertexAI and Kubeflow for efficient and scalable deployment pipelines. Integrate outputs with visualization platforms to deliver actionable insights and drive decision-making. Innovation: Stay current with trends in AI and data science, including LLMs, grounding techniques, and innovations in temporal and sequential data modeling. Regularly conduct POCs to experiment with emerging tools and technologies. Code Practices & Engineering: Write clean, maintainable, and scalable code following industry best practices. Adhere to version control (e.g., Git) for collaborative development and maintain coding standards. Implement error handling, logging, and monitoring to ensure reliability in production systems. Collaborate with other teams to integrate data science models into broader system architectures. Performance Optimization: Optimize model and data processing pipelines for computational efficiency and scalability. Use parallel processing, distributed computing, and hardware accelerators (e.g., GPUs, TPUs) where applicable. Documentation & Reusability: Maintain comprehensive technical documentation for all solutions. Design and build reusable assets to streamline future implementations. Technical Tools And Platforms Google Cloud (BigQuery, VertexAI, Cloud Storage) Python (Libraries: TensorFlow, Scikit-learn, XGBoost, LightGBM, etc.) SQL/NoSQL databases MLOps frameworks (Kubeflow, Vertex AI Pipelines) Visualization tools (Power BI, Tableau, Google Data Studio) Show more Show less

RecommendedJobs for You

Navi Mumbai, Maharashtra, India

Indore, Madhya Pradesh, India

Pune, Maharashtra, India

Mumbai, Maharashtra, India

Hyderabad, Telangana, India

Ahmedabad, Gujarat, India

Pune, Maharashtra, India