The CWs will support efforts to perform data mitigations on large scale datasets (image, video, text) leveraged by FAIR research teams. The goal is to proactively mitigate potential risks associated with these datasets. An ideal candidate is someone with a bachelor's degree in computer science from an accredited institution. Also someone who has worked with AI related tools in any capacity.
Job Responsibilities :
Preprocessing : converting original datasets into a format that can be consumed by mitigation pipelines.
Filtering : running filtering using Integrity's pipeline.
Implement the frontend (in React) and backend (in PHP or Python) of new features.
Post-processing : consuming filtering results to filter in the original datasets, repackaging, and re-ingestion.
Write and code logical and physical database descriptions and specify identifiers of database to management system or direct others in coding descriptions.
Optimization : identify optimization opportunities and improve the process.
Skills :
5+ years of relevant experience.
Software engineering skills include writing scripts to automate file processing and data transferring, and creating tools to improve productivity and streamline workflows.
Data Management : Data pipeline building. Data processing and cleaning, transformation and formatting, data quality control and validation.
Communication - effective communication skills to collaborate with stakeholders and team members.
Must-Have Skills :
5+ experience in Python, PHP and React a must.
Some data management experience, e.g. SQL, process large data.
Background in AI building tools.
Able to be flexible and work well in different environments with varying tasks and responsibilities.
Education / Experience :
Bachelor's degree in computer science, software engineering or relevant field required.