Job Description :
TECHNOGEN, Inc. is a Proven Leader in providing full IT Services, Software Development and Solutions for 15 years.



TECHNOGEN is a Small & Woman Owned Minority Business with GSA Advantage Certification. We have offices in VA; MD & Offshore development centers in India. We have successfully executed 100+ projects for clients ranging from small business and non-profits to Fortune 50 companies and federal, state and local agencies.


This is Tony from TechnoGen Inc. And I am writing to see whether you are interested in an exciting/challenging opportunity in Austin,  Texas.Kindly reach me on or  


Role : ETL Developer
Location : Austin,  Texas
Duration : Long Term  

Notes :Hadoop, Informatica BDM, CDC and ETL are must

The ETL Developer delivers full-stack data solutions across the entire data processing pipeline. This relies on systems engineering principles to design and implement solutions that span the data lifecycle to: collect, ingest, process, transform, store, persist, access, and deliver data at scale and at speed. It includes knowledge of local, distributed, and cloud-based technologies; data transformation and smart caching; and all security and authentication mechanisms required to protect the data.



Technical Skills:

·         Experience in Data Management, Data Integration & Analytics in diverse contexts with depth in data & information architectures - structured and unstructured preferably in Transportation or in large government for descriptive/diagnostic and predictive, prescriptive analytical needs.

·         Past or current experience in ETL development using Informatica or other compatible ETL solution

·         Expereince using Information BDM (Big Data Management)

·         Expertise in big data application data architecture and supporting implementation

·         Experience with one or more of the state of Texas government agencies

·         Knowledge of modern enterprise data architectures, design patterns, and data toolsets and the ability to apply them

·         Strong knowledge and experience of SQL and proficiency in data modeling techniques and understanding of normalization

·         Experience implementing or supporting Data Integration of Big Data with Sqoop or similar tools

·         Experience working with AWS S3 Storage, parquet and ORC files.

·         Strong problem solving, conceptualization, and communication skills

·         Understanding of enterprise service bus architectures and rest services using Kafka or compatible solution

·         Leveraging Big data & streaming technologies within AWS for data ingestion, transformation & persistence Sqoop, hive, kafka, nifi, oozie, java , Python, Spark, Hbase,Hartonworks Hadoop

·         Experience designing & managing Data Marts & Data Warehouse platforms.

·         Strong analytics & reporting skills experience with BI tools like Tableau, qlik, Ms Power BI Data modeling for data lake, data warehousing OLTP, normalized models,de-normalized, dimensional methods.

·         Strong data integration skills and experience especially around moving large data sets in batch & near real time across cloud & ground desired

·         Demonstrated success engaing business partners in a consultative manner and turning business concepts into well designed technology solutions.