Data scientist / engineer, looking to connect with organizations that need help making sense of a lot of data.
Strong software engineer who is passionate about large-scale distributed systems and cares about producing clean, elegant, maintainable, robust, well-tested Scala code.
Big Data / Core Skill: Apache Cassandra (DevOps, Data Modeling)
(Deep knowledge of Cassandra's internals and how to use it effectively.)
Big Data / Secondary Skill: Apache Flink, Spark Streaming
Big Data / Secondary Skill: Graph Modeling / Algorithms / Queries (with Spark GraphFrames and Neo4J)
Big Data / Other: Apache Kafka (incl. KafkaConnect / KafkaStreams / SchemaRegistry)
Programming Languages: Scala (highly proficient), Python (proficient)
Other Secondary Skills: Akka, Git, Docker, Apache Ignite, Scalding, Avro, Parquet, Zookeeper, HDFS, ElasticSearch, AWS (EC2/S3/RedShift/DataPipeline).
Other Secondary Skills: Machine Learning with Spark (Linear / Logistic Regression, Decision Trees, Naive Bayes, Alternating Least Squares / Recommender Systems, K-Means Clustering, Frequent Pattern Mining, TensorFlowOnSpark)
Professional Background (formerly): ETL Developer / Traditional DWHs / Kimball's and Data Vault Methodologies