
In Bearbeitung
Veröffentlicht
Bezahlt bei Lieferung
For my data‐management course homework, I must build a complete ETL workflow with any free or open-source tool such as Talend Open Studio, Apache NiFi, or Pentaho. The task starts with extracting a small, publicly available sample data set (CSV, JSON, or a simple relational dump), then cleaning and transforming it—deduplicating records, resolving missing or inconsistent values, and normalising key fields where needed. Once the data is tidy, it has to be loaded twice: first into a staging target (a plain relational table or file storage) and then into a basic star- or snowflake-style data-warehouse schema so I can run simple analytical queries afterward. I will need the full project files, transformation jobs, and a concise write-up that walks through each step, explains the design decisions, and shows the final row counts before and after every major operation. Screenshots or log excerpts that prove the pipeline runs end-to-end are essential. Acceptance criteria • All ETL jobs execute without errors from extract through warehouse load • Cleaning logic is clearly documented and reproducible • The warehouse tables populate with the expected record counts and data types • A short report (≈2 pages) summarises the process, tools used, and results, accompanied by any SQL used for validation Please keep tool choice flexible—as long as the solution is fully open-source and I can rebuild it locally, it meets the brief.
Projekt-ID: 40252217
15 Vorschläge
Remote Projekt
Aktiv vor 12 Tagen
Legen Sie Ihr Budget und Ihren Zeitrahmen fest
Für Ihre Arbeit bezahlt werden
Skizzieren Sie Ihren Vorschlag
Sie können sich kostenlos anmelden und auf Aufträge bieten

Hello I hope you're doing well. I understand you're looking for Build Sample ETL Pipeline Assignment I am the ideal candidate for your project. I have read the provided job description and I understand what you are looking for. I have over 10+ years of experience NoSQL Couch & Mongo, Pentaho, Hadoop, Data Warehousing, Elasticsearch, Data Extraction, ETL, Data Management .Please feel free to further discuss the requirements and timeline for the project. I'd be happy to assist you. I am ready to start right now. ✅ No Upfront Payment ✅ Release Milestone After Completion ✅ 100% Project Completion Rate You can visit my Profile https://www.freelancer.com/u/HiraMahmood4072 Thank you
$100 USD in 2 Tagen
4,6
4,6
15 Freelancer bieten im Durchschnitt $152 USD für diesen Auftrag

I’ll build a complete open-source ETL pipeline (Talend/NiFi/Pentaho), including extraction, cleaning (dedupe, null handling, normalization), staging load, and a star-schema warehouse with validated row counts and SQL checks. You’ll receive full project files, reproducible jobs, screenshots/log proof, and a concise 2-page report explaining design decisions and results.
$140 USD in 1 Tag
4,1
4,1

Hello Dear! I write to introduce myself. I'm Engineer Toriqul Islam. I was born and grew up in Bangladesh. I speak and write in English like native people. I am a B.S.C. Engineer of Computer Science & Engineering. I completed my graduation from Rajshahi University of Engineering & Technology ( RUET). I love to work on Web Design & Development project. Web Design & development: I am a full-stack web developer with more than 10 years of experience. My design Approach is Always Modern and simple, which attracts people towards it. I have built websites for a wide variety of industries. I have worked with a lot of companies and built astonishing websites. All Clients have good reviews about me. Client Satisfaction is my first Priority. Technologies We Use: Custom Websites Development Using ======>Full Stack Development. 1. HTML5 2. CSS3 3. Bootstrap4 4. jQuery 5. JavaScript 6. Angular JS 7. React JS 8. Node JS 9. WordPress 10. PHP 11. Ruby on Rails 12. MYSQL 13. Laravel 14. .Net 15. CodeIgniter 16. React Native 17. SQL / MySQL 18. Mobile app development 19. Python 20. MongoDB What you'll get? • Fully Responsive Website on All Devices • Reusable Components • Quick response • Clean, tested and documented code • Completely met deadlines and requirements • Clear communication You are cordially welcome to discuss your project. Thank You! Best Regards, Toriqul Islam
$100 USD in 3 Tagen
3,7
3,7

Hello. Thanks for your job posting. ⭐I'm the developer you're looking for.⭐ I can successfully complete your project. Let's chat for a more detailed discussion. Thank you. Maxim
$140 USD in 7 Tagen
3,1
3,1

I will build a complete ETL workflow using a free or open-source tool such as Talend Open Studio, Apache NiFi, or Pentaho, extracting a small publicly available sample data set, cleaning and transforming it, and loading it into a staging target and a basic star- or snowflake-style data-warehouse schema, providing full project files, transformation jobs, and a concise write-up, meeting the acceptance criteria and adapting to the proposed budget. Waiting for your response in chat! Best Regards.
$146 USD in 3 Tagen
0,0
0,0

Hello, Hope you are doing well — I can build your complete open-source ETL workflow using Talend Open Studio or Apache NiFi (your choice), extracting a public CSV/JSON dataset, performing reproducible cleaning steps (deduplication, null handling, normalization), loading first into a staging relational table (PostgreSQL), then into a properly designed star-schema warehouse with documented transformations and validated row counts at each stage. I will provide full project files, transformation jobs, SQL validation scripts, execution logs/screenshots proving end-to-end success, and a concise ~2-page technical report clearly explaining architecture decisions, schema design, and before/after record metrics. I recently completed a university ETL project involving public census data transformed into a dimensional warehouse for analytical querying, and I have 10 years of experience in data engineering and database systems. Do you have a preferred dataset domain (e.g., healthcare, retail, transportation), and must the warehouse follow a strict star schema or is a snowflake design acceptable if normalization improves clarity? Should the final environment be fully containerized (e.g., Docker with PostgreSQL) for easier local rebuild and grading consistency? Thank you. Maksim
$205 USD in 7 Tagen
0,0
0,0

Hi, Our devs looked at your project and noticed a potential bottleneck in the current system's architecture. It seems like scaling efficiently might be a challenge due to the data-heavy operations you're running. Our backend lead has experience in architecting systems that handle high throughput, ensuring smooth scalability. We recently completed a project for a fintech company where we developed a robust system using Node.js and AWS, handling over a million transactions daily without any downtime. This experience aligns perfectly with your needs, ensuring reliability and efficiency. I'll be your direct technical point of contact, and we'll set up a staging environment for testing and feedback. This way, you can have real-time updates and insights into our process. How do you envision the long-term growth of your platform? Excited to explore how we can support your vision.
$100 USD in 7 Tagen
0,1
0,1

Hello, I hope you are doing well. I’m a data engineer with hands-on experience building end-to-end ETL pipelines using open-source tools such as Apache NiFi, Talend Open Studio alternatives, and PostgreSQL. I can craft a complete, reproducible workflow that extracts a public dataset in CSV/JSON, cleans and normalizes it (deduplication, missing values handling, standardizing key fields), and loads it into a staging area and then into a star- or snowflake-style warehouse. I’ll deliver the full project artifacts: transformation jobs, data-model definitions, validation SQL, and a concise write-up that explains design decisions and shows final row counts at each major step, supported by screenshots or logs proving end-to-end execution. The solution will be fully open-source and locally rebuildable, with clear, reproducible steps and documentation to match your course requirements. Best regards, Billy Bryan
$250 USD in 5 Tagen
0,0
0,0

Hi there, Regarding your project, a common oversight is the inefficient handling of data processing tasks, leading to performance lags. My approach ensures optimal resource allocation and delivers a seamless user experience by intelligently prioritizing processes based on real-time demand. I've tackled similar challenges before and can apply this expertise to your needs. In a recent project, I streamlined a client's data pipeline, boosting their processing speed by 30%, which significantly improved user satisfaction and engagement. I include 30 days of post-deployment bug-fixing to ensure stability. What's the current data throughput, and how flexible is your existing system architecture for integration? Let's discuss how I can bring this expertise to your project.
$100 USD in 7 Tagen
0,0
0,0

As an experienced software test engineer and workflow automation specialist, I can assure you that I have the skills required to successfully fulfill your ETL project needs. My expertise in designing and implementing automation frameworks, including workflow automation with tools like n8n, Make, Zapier, Zoho, GoHighLevel, can be directly applied to streamline and automate your data management processes. Having worked with Pentaho and Talend Open Studio, I am familiar with the most effective ways to extract data from various sources like CSV and JSON, transforming and cleaning it as per the specific requirements. Additionally, my proficient knowledge in API testing and UI automation using tools like Selenium, Cypress and Playwright would enable me to thoroughly validate the data integration between your staging target and data warehouse schema. The ability to deliver large-scale data extraction using Scrapy, Selenium, Python, BeautifulSoup etc has given me a thorough understanding of how data transformations require a keen attention to detail combined with efficient tools to maintain integrity. Ultimately, my goal is to deliver a complete ETL workflow that is error-free from extraction through warehouse load - a task I have excelled in throughout the 10 years in my career.
$140 USD in 7 Tagen
0,0
0,0

Greetings I can build a complete ETL workflow for your data-management course homework using a free, open-source tool such as Talend Open Studio, Apache NiFi, or Pentaho. The workflow will start with extracting a small, publicly available sample data set (CSV, JSON, or relational dump), then proceed to clean and transform the data by deduplicating records, resolving missing or inconsistent values, and normalizing key fields where necessary. Once cleaned, the data will be loaded twice: first into a staging target (relational table or file storage) and then into a star- or snowflake-style data warehouse schema suitable for running analytical queries. I will provide all project files, transformation jobs, and a concise write-up explaining each step, design decisions, and showing row counts before and after every major operation. Screenshots or log excerpts will demonstrate that the pipeline runs end-to-end successfully. The workflow will meet the following criteria: • ETL jobs execute without errors from extraction through warehouse load • Cleaning logic is fully documented and reproducible • Warehouse tables are populated with the expected record counts and correct data types • A short report (≈2 pages) summarises the process, tools used, results, and includes any SQL queries used for validation Could you please share the sample dataset you want to use and any specific schema requirements so I can start building the ETL workflow? Regards, MH.
$140 USD in 3 Tagen
0,0
0,0

I can build a complete end-to-end ETL workflow using a fully open-source tool such as Talend Open Studio, Apache NiFi, or Pentaho (based on your preference). The project will include data extraction from a public dataset, cleaning and transformation steps (deduplication, missing value handling, normalization), staging load, and final loading into a star or snowflake data warehouse schema. I will provide fully documented ETL jobs, project files, SQL validation scripts, and a concise report explaining each step, design decisions, and record counts before and after transformations. Screenshots and execution logs will be included to demonstrate successful pipeline runs from extraction to warehouse loading. The final solution will be structured so you can easily rebuild and run it locally without proprietary software. I focus on clear, reproducible data workflows and clean documentation to ensure the project meets academic and technical acceptance criteria.
$100 USD in 5 Tagen
0,0
0,0

Riyadh, Saudi Arabia
Zahlungsmethode verifiziert
Mitglied seit Sept. 20, 2025
$8-15 USD / Stunde
$300-450 USD
$8-15 USD / Stunde
$250-750 USD
$30-250 USD
$10-30 CAD
€250-750 EUR
$15-25 USD / Stunde
₹1500-12500 INR
₹12500-37500 INR
$25-50 USD / Stunde
₹12500-37500 INR
$15-25 USD / Stunde
$10-30 USD
₹750-1250 INR / Stunde
₹600-1200 INR
$1500-3000 USD
$20000-50000 CAD
€12-18 EUR / Stunde
$30-250 USD
₹600-1500 INR
₹1500-12500 INR
₹750-1250 INR / Stunde