read our blogs

Blogs or Expert Columns

Taming of the Bots – An Insider’s Story – by Dj Das

Taming of the Bots - An Insider's Story Dj Das will be presenting this topic at the Global Data Science Conference. Enterprises have a new friend that saves them time, efforts and money. Bots are being adopted by enterprises in all aspects of their daily business needs. Expectations are wild and sometimes unrealistic, as to how it happens with every new technology. This session will focus on how to make the Bots work for enterprises, the does and don’ts, the gotchas and oops - irrespective of the technology platform used for building and deploying these Bots. Check out the complete agenda here. [...]

Global Data Science Conference – Free Passes!

ThirdEye is a proud sponsor of Global Data Science Conference that is being held on April 2nd - 4th, 2018 on all industry verticals. The Conference allows practitioners to discuss Data Science through effective use of various data management techniques. This will be a good opportunity to network with attendees and Speakers. Location: Santa Clara Convention Center, 5001 Great America Parkway, Santa Clara, CA 95054 Agenda:  http://www.globalbigdataconference.com/santa-clara/global-data-science-conference/schedule-98.html Confirmed Keynote Speakers include: Deep Learning Article Embedding for Fake News Evaluation (Mike Tamir, Head of Data Science, Uber) TensorFlow Lite: How To Accelerate Your Android And IOS App With AI (Kaz Sato, Lead Developer Evangelist, [...]

ODSC East – Discounted Pricing!

ODSC East - Boston - May 1-4 2018 ThirdEye is pleased to announce its partnership with Open Data Science Conference (ODSC) and announce the ODSC East event in Boston fromMay 1st through May 4th 2018. This year’s 2018 ODSC East brings together the most influential data scientists, practitioners, innovators, and thought leaders in Data Science, Artificial Intelligence, and Big Data. ODSC East showcases the best and the brightest when it comes to Data Science, including many open source data science pioneers. This 4-day event will feature a total of 202 sessions: 93 talks, 63 hands-on workshops, and 36 intense [...]

RedisConf18 – Free Passes!

ThirdEye is pleased to announce RedisConf18 that is happening in SF next month. ThirdEye has some passes to giveaway. Use promo code 'SVBIGDATACLOUD' to register for a FREE conference pass at http://redisconf.com. Here are some of the awesome talks and speakers: BigData, Analytics & Machine Learning (across both days) Redis Streams - Salvatore Sanfillipo, Creator of Redis Video Experience Operational Insights in Real Time – Aditya Vaidya, Oath Inc Large Data Pipelines - Piyush Verma, Oogway.in Time-Series Data in Real-time - TimeSeries Module, Danni Moiseyev, Redis Labs Redis Analytics Use Cases - Leena Joshi, Kyle Davis, Redis Labs 25 Billion Messages Per Day - Jongyeol Choi, LINE+ Corporation Real-time log analytics using [...]

Chat with TEbot – the Big Data Advisor!

ThirdEye Data has launched Syra, a Custom Enterprise Chatbot Development service. As part of the service, we are launching TEbot today, a Big Data Advisor. TEbot answers your questions about Big Data technologies to help you decide on which technologies to use when, why, and how. Chatbots increase site visitors engagement & retention, which results in higher sales conversions. Reach out to us to build your Custom Enterprise Chatbot and watch your business grow! Wanna Build a Chatbot? Related Blogs: ThirdEye Data Ranked Globally as a Top Big Data, BI and B2B Company ThirdEye Data Ranked as Top 3 Big Data Analytics Company ThirdEye Data [...]

Bringing Big Data to the Masses

A true leader does not follow trends, he initiates them. Dj Das, CEO, Third Eye Consulting Services and Solutions personifies this fact. Armed with strong foresight that Big Data technologies would be the next frontier for innovation, competition, and productivity, he took up the charge to fuel the Big Data fire in the tech world. The world took some time but finally came up to speed with Dj's strong belief in Big Data technologies being the future. Today he is a known evangelist for Big Data; runs a much followed meetup & blog named BigDataCloud, (a not for profit [...]

Theano

Overview Theano is a Python library for fast numerical computation that can be run on the CPU or GPU. It is a key foundational library for Deep Learning in Python that you can use directly to create Deep Learning models or wrapper libraries that greatly simplify the process. In this post, you will discover the Theano Python library. What is Theano? Theano is an open source project released under the BSD license and was developed by the LISA (now MILA) group at the University of Montreal, Quebec, Canada (home of Yoshua Bengio). It is named after a Greek mathematician. At its heart, [...]

Caffe 

Caffe is a deep learning framework made with expression, speed, and modularity in mind. It is developed by Berkeley AI Research (BAIR) and by community contributors. Yangqing Jia created the project during his PhD at UC Berkeley. Caffe is released under the BSD 2-Clause license. Why Caffe? Expressive architecture encourages application and innovation. Models and optimization are defined by configuration without hard-coding. Switch between CPU and GPU by setting a single flag to train on a GPU machine then deploy to commodity clusters or mobile devices. Extensible code fosters active development. In Caffe’s first year, it has been forked by over [...]

Teradata

What is Teradata? Teradata is an RDBMS (relational database management system) which includes features: It’s built on completely parallel architecture which means single task will be divided into smaller chunks and compute simultaneously hence faster execution. Teradata system is a shared-nothing architecture in which each node is independent and self-sufficient. Also, each logical processor (AMP) is responsible only for their own portion of the database. Supporting industry standard ANSI SQL to communicate with Teradata. Teradata database can be accessed by multiple concurrent users from different client applications via popular TCP/IP connection or IBM mainframe channel connection. Why use Teradata? [...]

Spark ML

Spark ML Spark.ml is a new package introduced in Spark 1.2, which aims to provide a uniform set of high-level APIs that help users create and tune practical machine learning pipelines. It is currently an alpha component, and we would like to hear back from the community about how it fits real-world use cases and how it could be improved.Note that we will keep supporting and adding features to spark.mllib along with the development of spark.ml. Users should be comfortable using the spark.mllib features and expect more features coming. Developers should contribute new algorithms to spark.mllib and can optionally [...]

Java

Java is a programming language and a computing platform for application development. It was first released by Sun Microsystem in 1995 and later acquired by Oracle Corporation. It is one of the most used programming languages. Oracle has two products that implement Java Platform Standard Edition (Java SE) 8: Java SE Development Kit (JDK) 8 and Java SE Runtime Environment (JRE) 8. JDK 8 is a superset of JRE 8 and contains everything that is in JRE 8, plus tools such as the compilers and debuggers necessary for developing applets and applications. JRE 8 provides the libraries, the Java [...]

ETL – Extract Transform Load

Overview One of the ways to integrate with external systems is using the Extract, Transform, Load (ETL) system. The ETL system creates databases which can be accessed directly by 3rd party tools and solutions. It also allows scheduled execution of transformation scripts on the Data Center Operation server. Together this gives the ability to extract and enter data on the Data Center Operation server. Based on the ETL system, it is possible to develop custom solutions, integrating DCO with a broad range of data sources. ETL can be used in 2 ways: Importing Data in to and Exporting Data from Data Center Operation External system integration using ETL ETL modeling concepts The general framework for ETL [...]

Multi-tenancy

Multi-tenancy is an architecture in which a single instance of a software application serves multiple customers. Each customer is called a tenant. Tenants may be given the ability to customize some parts of the application, such as the color of the user interface (UI) or business rules, but they cannot customize the application's code. Multi-tenancy can be economical because software development and maintenance costs are shared. It can be contrasted with single-tenancy, an architecture in which each customer has their own software instance and may be given access to the code. With a multi-tenancy architecture, the provider only has [...]

Azure HDInsight Spark

Azure HDInsight Spark Overview Introduction to Spark on HDInsight This article provides you with an introduction to Spark on HDInsight. Apache Spark is an open-source parallel processing framework that supports in-memory processing to boost the performance of big-data analytic applications. Spark cluster on HDInsight is compatible with Azure Storage (WASB) as well as Azure Data Lake Store. Hence, your existing data stored in Azure can easily be processed via a Spark cluster. When you create a Spark cluster on HDInsight, you create Azure compute resources with Spark installed and configured. It only takes about 10 minutes to create a Spark cluster [...]