Machine Learning Engineer
Acuity
- Virginia
- $6,000 per month
- Permanent
- Full-time
- Shall create data science and data engineering products that include data models, data ingestion/transform, analytics which may include machine learning, and some form of output as either a machine-readable format (e.g. file output, database output, standard/streaming output) or a user interface or dashboard.
- Shall develop robust data engineering pipelines utilizing Apache Nifi and Python to include use and development of REST APIs and microservices.
- Shall clean, parse and transform data from multiple file types into appropriate database architectures (e.g. SQL, noSQL, and graph) which perform at scale.
- Shall develop, deploy and provide feature enhancements using Python for data science products and services, to include Python packages, code documentation, notebooks, and microservices.
- Shall ensure that all technical development and content complies with Client’s security policies and regulations.
- Shall work closely with the Client to review and track data science and data engineering requirements and provide regular updates to clearly explain the project status and results in both written and verbal or multimedia briefings.
- Shall consult with stakeholders to determine present and future user needs for Client consideration.
- Shall be required to communicate and collaborate across organizational boundaries, to include other staff and contractor teams.
- Shall work with Client staff and personnel, as well as external stakeholders on mission-based projects.
- Shall implement data science and data engineering requirements as defined by the Client.
- Shall determine how requirements are satisfied. Project priorities are managed by the staff manager of the business unit. Planned activities shall be coordinated with all stakeholders and approved by the Client.
- Shall utilize project management systems such as Confluence and JIRA to define and track requirements.
- Shall ensure that technical solutions leverage industry best practices, designing for security and excellence while minimizing the total cost of ownership.
- Demonstrated on-the-job experience developing Python programming packages to include REST APIs and microservices.
- Demonstrated experience developing and testing reusable Python code.
- Demonstrated experience with machine learning techniques including natural language processing and computer vision.
- Demonstrated on-the-job experience using Linux flavored operating systems (OS) and with automating work flows using Bash scripting.
- Demonstrated experience using data engineering tools such as Apache Nifi to preprocess, modify, aggregate, load, index, and archive large data collections.
- Demonstrated experience performing the extraction, transformation, and loading (ETL) of structured and unstructured data into pipelines to ensure it is ingested into downstream systems with accuracy, reliability, and consistency at scale.
- Demonstrated experience modeling, structuring, cleaning, and conditioning data from multiple sources and in multiple different formats, languages, and encodings.
- Demonstrated experience using Elasticsearch and Kibana technologies.
- Demonstrated experience using code repositories such as Git.
- Demonstrated experience developing robust documentation for code, Python packages and data science methodologies.
- Demonstrated experience explaining complex technical issues to more junior data scientists, in graphical, verbal, or written formats.
- Demonstrated experience delivering results to stakeholders through written documentation and oral briefings.
- Demonstrated experience working with multiple stakeholders.
- Demonstrated experience with Nutanix or NetApps.
- Demonstrated experience with cloud services, such as AWS.
- Demonstrated experience using big data processing tools such as Apache Spark or Trino.
- Demonstrated experience using container frameworks such as Docker or Kubernetes.
- Demonstrated experience using data visualizations tools such as Tableau.
- TS/SCI w/Poly