COVID has dramatically altered all of our lives. The corresponding flurry of information — and sometimes misinformation — has made these changes more acute and harder to unpack.
To empower citizens, journalists, and policymakers, Monument has assembled below easy-to-follow instructions for building your own machine learning models for COVID’s spread, mortality rate, and other important factors.
We have laid out the instructions in such a way to make them easy to modify for your own analyses. As always, if you have any questions, feel free to contact us at email@example.com.
The specific questions we’re answering here are:
NEW YORK, N.Y. — Monument is pleased to announce the roll out of “model serving” capabilities. Model serving enables users to train algorithms on historical data, save the parameterized model, and deploy this model on new data that the model has not previously seen.
Model serving complements the “training” and “validation” steps in the Machine Learning workflow. It is used for two principal workflows: “testing” and “live:”
Stop putting prices on things by guessing. You’re leaving money on the table. A great use-case for Machine Learning is building pricing engines. No coding or data science expertise required.
Whether you’re buying or selling a product — any product! — it’s helpful to use Machine Learning to get a sense of whether a given price is too high or too low. This is true for any product, in particular commodities like computers, cars, steel, corn, fertilizer, airline tickets, concert tickets — whatever.
Pop quiz. In the above chart of eleven historical datapoints, which of the two lines in the picture above would be the most accurate for predicting the location of a new, twelfth datapoint? The blue line seems great — it’s fit perfectly to your known historical datapoints! The diagonal black line seems “dumb” in comparison, as it only manages to intersect three out of the eleven historical datapoints.
It may be surprising to some readers that the diagonal black line may predict the location of the next datapoint more accurately than the blue line.
Deciding how to choose between these models requires you to understand the concept of overfitting. Recognizing and combating overfitting is a core task in Machine Learning. There is no “magic bullet” approach to dealing with overfitting that will work in all use-cases, so it’s useful to gain some intuition. …
Data science does not have to mean months or years of work to get started. With a no-code platform you can go from literally zero coding or data science experience to running your first algorithm in minutes — literally!
It’s hard to find the meaning in arrays of numbers, and visualizing your data is its own coding hurdle. No-code tools automatically visualize your results so that you can see the results of your work.
Regressions, Neural Networks, Kalman Filters. If you’re learning to code with, for example, Python, you’re going to deal with an entirely new learning curve for each algorithm you try. This will make it hard to compare models and learn by these comparisons. …
NEW YORK, N.Y. — The Monument team is pleased to announce the addition of classification and regression methods into our zero-code machine intelligence platform. Users can now estimate any types of classes or levels for your data. Monument will automatically adjust to your data set.
These classification and regression algorithms open up entirely new use cases across a variety of industries, such as fraud detection, value appraisal, customer categorization, and machine failure.
The classification algorithms are Support Vector Machine (SVM), Light Gradient Boosted Machine (LightGBM), and Logistic Regression (LogReg). These features expand on Monument’s robust suite of time-series algorithms.
Classification algorithms enable users to use training datasets to build algorithms that can infer characteristics about new data. Typical use cases…
Predicting the likelihood of future events like fraud or payment defaults is a classic use-case of machine learning. With its drag-and-drop interface, Monument makes it easier than ever to tackle this classification problem. In this tutorial, in a matter of minutes, we will use real-world data on credit card default probability to train an algorithm to detect payment defaults.
In the “Data Folder” of the University of California-Irvine repository linked above, there is a file called
default of credit card clients.xls. When we open it up in a spreadsheet, it looks like this:
Monument (www.monument.ai) enables you to quickly apply algorithms to data in a no-code interface. But, after you drag the algorithms onto data to generate predictions, you need to decide which algorithm or combination of algorithms is most reliable for your task.
In the ocean temperature tutorial, we cleaned open remote sensing data and fed the data into Monument in order to forecast future ocean temperatures. In that case, we used visual inspection to evaluate the accuracy of different algorithms, which was possible because the historical data roughly formed a sine curve. …
In this tutorial, we’re going to show you how to take open source data from the National Oceanic and Atmospheric Administration (NOAA), clean it, and forecast future temperatures using no-code machine learning methods.
This particular data comes from the Harmful Algal BloomS Observation System (HABSOS). There are several interesting questions to ask of this data — namely, what is the relationship between algal blooms and water temperature fluctuations. For this tutorial, we’re going to start with a basic question: can we predict what temperatures will be over the next five months?
The first part of this tutorial deals with acquiring and cleaning the dataset. There are a lot of approaches to this; what is shown below is just one approach. Further, if your dataset is already clean, you can skip all that “data engineering” and jump straight into no-code AI bliss…
NEW YORK, NY. — Monument announces the introduction of three new algorithms and an integration with Binance’s API for cryptocurrency prices. These additions enable cryptocurrency traders to build their own trading algorithms with greater ease than ever before.
Previously, cryptocurrency traders without data science expertise faced two roadblocks when trying to find and use investment signals: organizing historical data and applying complex algorithmic methods. As a result, the majority of the cryptocurrency trading community relied on simple data and widely-available signals.
Monument changes the status quo. The platform’s new features allow any user to easily pull data from Binance and apply a suite of industry-leading algorithms in a simple drag-and-drop interface. An early user reports, “Being able to construct my own trading bots to find hidden trading signals has freed me from reliance on buying out-of-the-box trading bots.” …