DYI Rain Prediction Using Arduino, Python and Keras
by danionescu in Circuits > Arduino
4785 Views, 11 Favorites, 0 Comments
DYI Rain Prediction Using Arduino, Python and Keras
First a few words about this project the motivation, the technologies involved and the end product that we're going to build.
So the big aim here is obviously to predict the rain in the future (we'll try 6 hours). The prediction will be a yes or no (boolean in programming terms). I've searched from tutorials on this matter and i haven't found one that is complete in every way. So mine will take this to a whole new approach and get into every aspect of it. For that we're going to:
- build the weather station our selves. The station should be completely off grid with a solar panel, and a extremely low power mode (a few dozens microamps hour)
- program the station so it gather data and transmits it every ten minutes to a base station
- collect the data on the base station and store it (in a database)
- using neural networks (Keras library) and other python libraries like pandas filter, clean and preprocess the data, then feed it to a neural network to train a "model" for predicting that it shall or not rain
- finally predict whether or not it will rain on the next 6 hours and notify users via email
I've personally used this weather station to collect data (you can download the data in the next steps if you wish). With only about 600 days of weather data the system can make a prediction if it will rain or not in the next 6 hours with an accuracy of around 80% depending on the parameters which is not that bad.
This tutorial we'll take you through all the necessary steps to predict rainfall from scratch. We'll make an end product that does a practical job, without using external weather or machine learning API-s. We'll learn along the way how to build a practical weather station (low power and off the grid) that really collects data over long periods of time without maintenance. After that you'll learn how to program it using Arduino IDE. How to collect data into a database on a base station (server). And how to process the data (Pandas) and apply neural networks (Keras) and then predict the rainfall.
Parts and Tools for Building the Station
Parts:
1. Small plastic box with detachable lids (mine has screws). The box size should be big enough to fit the small components and batteries. My box has 11 x 7 x 5 cm
2. three AAA battery holder
3. three AAA rechargeable batteries
4. 6V small solar panel
5. Arduino pro mini 328p
6. a diode, 1N4004 (to prevent reverse current from the batteries to the panel)
7. a small NPN transistor and 1k resistor (for switching power to the components on and off)
8. a rain sensor
9. HC-12 serial communication module
10. HC-12 USB serial module (for the base station)
11. BME280 bosch sensor module (for humidity, temperature, pressure)
12. BH1750 light sensing module
13. PCB, wires, solder, KF301-2P plug in screw connector, male&female pcb connectors, glue
14. 3.3V regulator
15. a base station: a PC, or a development board running all the time. It's role it's gathering data, training the rain prediction model and making predictions
Tools:
1. USB to serial FTDI adapter FT232RL to program the Arduino pro mini
2. Arduino IDE
3. Drill
4. Fine blade saw
5. Screwdrivers
6. Soldering iron
7. Wire cutter
Skills:
1. Soldering, check this tutorial
2. Basic arduino programming
3. Linux service configuration, package installation
4. Some programming skills
Building the Weather Station
The weather station is composed from the following sets of components:
1. the box with the solar panel glued to it
2. the PCB with the electronics inside
3. the battery holder also inside
4. the BME280 and light and rain sensors on the outside
1. The box needs 4 holes, one for the solar panel wires, other three for the sensors that will be placed on the outside. First drill the wholes, they should be large enough for the male-female wires to stick out and go to the sensors. After the wholes are drilled, glue the panel to one side of the box and get it's wires through a hole inside
2. The PCB will hold the arduino, HC-12, 3.3V regulator, diode, transistor, resistor and two KF301-2P
- first solder the two female PCB connectors on the PCB for the arduino, solder the male pcb connectors to the arduino and place the arduino on the PCB
- the arduino led should be removed or at least one of it's pin. this is very important because the led will draw a large amount of power. Be careful not to damage other components
- solder the transistor, resistor, and 3.3V regulator
- solder the two KF301-2P. One will be for the solar panel, the other for the battery holder
- solder three female PCB connectors: for the light sensor, BME280 and rain sensor
- solder small wires to connect all the PCB components (check the pictures and the fritzing schamatic)
3. place 3 charged AAA NiMH batteries inside the holder and place it inside the box, connecting the wires into the KF301-2P connector
4. connect the BME280 and light sensors from outside the box to their corresponding male connectors
For the rain sensor, solder three wires (Gnd, Vcc, signal) to it, and to the other side solder male pins that will go inside the box to its corresponding male connectors
The final thing it would be to place the station to it's final position. I've chose a position that is sheltered from the rain and snow. I've chosen longer wires for the rain sensor and placed it separately in the rain on a steady support. For the main box i've chosen a special kind of tape with adhesive (check the pictures) but anything that holds the box will do.
Downloads
Arduino Code
In this step you'll learn what external libraries are needed, we're overview the code and how it works and of course you'll get to download it or copy-paste it into Arduino IDE and upload it to the weather station.
The role of the weather station is to transmit to a base station every 10 minutes data about it's sensors.
First let's describe what the weather station program does:
1. read sensors data (humidity, temperature, pressure, rain, light, voltage)
2. transmits the encoded data through a second software serial line.
The encoded data looks like this:
H1:78|T1:12|PS1:1022|L1:500|R1:0|V1:4010|
The statement from above will mean that: humidity from station "1" is 78 percent, temperature from station 1 is 12 degrees, pressure is 1022 bars, light level 500 lux, rain 0, and voltage 4010 millivolts
3. power off the auxiliary components: sensors and communication device
4. puts the arduino in sleep mode for 10 minutes (this will make it consume less 50 microamps)
5. turn on the components and repeat steps 1 - 4
One little extra tweak here, if the voltage level is above 4.2 V the arduino will use the normal sleep function "delay(milliseconds)". This will greatly increase the power consumption and decreasing the voltage rapidly. This effectively prevents the solar panel from overcharging the batteries.
You can get the code from my Github repository here: https://github.com/danionescu0/home-automation/tre...
Or copy-pasting it from below, either way just remove the line with "transmitSenzorData("V", sensors.voltage);"
#include "LowPower.h"<br>#include "SoftwareSerial.h" #include "Wire.h" #include "Adafruit_Sensor.h" #include "Adafruit_BME280.h" #include "BH1750.h" SoftwareSerial serialComm(4, 5); // RX, TX Adafruit_BME280 bme; BH1750 lightMeter; const byte rainPin = A0; byte sensorsCode = 1; /** * voltage level that will pun the microcontroller in deep sleep instead of regular sleep */ int voltageDeepSleepThreshold = 4200; const byte peripherialsPowerPin = 6; char buffer[] = {' ',' ',' ',' ',' ',' ',' '}; struct sensorData { byte humidity; int temperature; byte rain; int pressure; long voltage; int light; }; sensorData sensors; void setup() { Serial.begin(9600); serialComm.begin(9600); pinMode(peripherialsPowerPin, OUTPUT); digitalWrite(peripherialsPowerPin, HIGH); delay(500); if (!bme.begin()) { Serial.println("Could not find a valid BME280 sensor, check wiring!"); while (1) { customSleep(100); } } Serial.println("Initialization finished succesfully"); delay(50); digitalWrite(peripherialsPowerPin, HIGH); } void loop() { updateSenzors(); transmitData(); customSleep(75); } void updateSenzors() { bme.begin(); lightMeter.begin(); delay(300); sensors.temperature = bme.readTemperature(); sensors.pressure = bme.readPressure() / 100.0F; sensors.humidity = bme.readHumidity(); sensors.light = lightMeter.readLightLevel(); sensors.voltage = readVcc(); sensors.rain = readRain(); } void transmitData() { emptyIncommingSerialBuffer(); Serial.print("Temp:");Serial.println(sensors.temperature); Serial.print("Humid:");Serial.println(sensors.humidity); Serial.print("Pressure:");Serial.println(sensors.pressure); Serial.print("Light:");Serial.println(sensors.light); Serial.print("Voltage:");Serial.println(sensors.voltage); Serial.print("Rain:");Serial.println(sensors.rain); transmitSenzorData("T", sensors.temperature); transmitSenzorData("H", sensors.humidity); transmitSenzorData("PS", sensors.pressure); transmitSenzorData("L", sensors.light); transmitSenzorData("V", sensors.voltage); transmitSenzorData("R", sensors.rain); } void emptyIncommingSerialBuffer() { while (serialComm.available() > 0) { serialComm.read(); delay(5); } } void transmitSenzorData(String type, int value) { serialComm.print(type); serialComm.print(sensorsCode); serialComm.print(":"); serialComm.print(value); serialComm.print("|"); delay(50); } void customSleep(long eightSecondCycles) { if (sensors.voltage > voltageDeepSleepThreshold) { delay(eightSecondCycles * 8000); return; } digitalWrite(peripherialsPowerPin, LOW); for (int i = 0; i < eightSecondCycles; i++) { LowPower.powerDown(SLEEP_8S, ADC_OFF, BOD_OFF); } digitalWrite(peripherialsPowerPin, HIGH); delay(500); } byte readRain() { byte level = analogRead(rainPin); return map(level, 0, 1023, 0, 100); } long readVcc() { // Read 1.1V reference against AVcc // set the reference to Vcc and the measurement to the internal 1.1V reference #if defined(__AVR_ATmega32U4__) || defined(__AVR_ATmega1280__) || defined(__AVR_ATmega2560__) ADMUX = _BV(REFS0) | _BV(MUX4) | _BV(MUX3) | _BV(MUX2) | _BV(MUX1); #elif defined (__AVR_ATtiny24__) || defined(__AVR_ATtiny44__) || defined(__AVR_ATtiny84__) ADMUX = _BV(MUX5) | _BV(MUX0); #elif defined (__AVR_ATtiny25__) || defined(__AVR_ATtiny45__) || defined(__AVR_ATtiny85__) ADMUX = _BV(MUX3) | _BV(MUX2); #else ADMUX = _BV(REFS0) | _BV(MUX3) | _BV(MUX2) | _BV(MUX1); #endif delay(2); // Wait for Vref to settle ADCSRA |= _BV(ADSC); // Start conversion while (bit_is_set(ADCSRA,ADSC)); // measuring uint8_t low = ADCL; // must read ADCL first - it then locks ADCH uint8_t high = ADCH; // unlocks both long result = (high<<8) | low; result = 1125300L / result; // Calculate Vcc (in mV); 1125300 = 1.1*1023*1000 return result; // Vcc in millivolts }
Before uploading the code, download and install the following arduino libraries:
* BH1750 library: https://github.com/claws/BH1750
* LowPower library: https://github.com/rocketscream/Low-Power
* Adafruit Sensor library: https://github.com/adafruit/Adafruit_Sensor
* Adafruit BME280 library: https://github.com/adafruit/Adafruit_Sensor
If you don't know how to do that, check out this tutorial.
Preparing the Base Station
The base station will consist of a linux computer (desktop, laptop or development board) with the HC-12 USB module attached. The computer must remain always on to collect data every 10 minutes from the station.
I've used my laptop with Ubuntu 18.
The installation steps:
1. Install anaconda. Anaconda is a python package manager and it will make it easy for us to work with the same dependencies. We'll be able to control the python version, and each package version
If you don't know how to install it check out this: https://www.digitalocean.com/community/tutorials/h... tutorial and follow the steps 1 - 8
2. Install mongoDb. MongoDb will be our main database for this project. It will store data about all the sensors time series. It's schemaless and for our purpose it's easy to use.
For installation steps check out their page: https://docs.mongodb.com/v3.4/tutorial/install-mon...
I've used an older version of mongoDb 3.4.6, if you follow the tutorial above you'll get exactly that. In principle it should work with the latest version.
[Optionally] add an index on date field:
mongo
use weather db.weather_station.createIndex({"date" : 1})
3. Download the project from here: https://github.com/danionescu0/home-automation . We'll be using the weather-predict folder
sudo apt-get install git git clone https://github.com/danionescu0/home-automation.gi...
4. Create and configure anaconda environment:
cd weather-predict # create anaconda environment named "weather" with python 3.6.2 conda create --name weather python=3.6.2 # activate environment conda activate weather # install all packages pip install -r requirements.txt
This will create an new anaconda environment and install the needed packages. Some of the packages are:
Keras (high level neural network layer, with this library we'll make all our neural network predictions)
pandas (useful tool that manipulates data, we'll use it heavily)
pymongo (python mongoDb driver)
sklearn (data mining and data analysing tools)
Configure the project
The config file is situated in the weather-predict folder and is named config.py
1. if you install MongoDb remotely or on a different port change the "host" or "port" in the
mongodb = { 'host': 'localhost', 'port': 27017 } ...
2. Now we need to attach the HC-12 USB serial adapter. Before than run:
ls -l /dev/tty*
and you should get a list of mounted devices.
Now insert the HC-12 in a USB port and run the same command again.It should be one new entry in that list, our serial adapter. Now change the adapter port in the config if necessary
serial = { 'port': '/dev/ttyUSB0', 'baud_rate': 9600 }
The other config entries are some file default paths, no need for a change there.
Use the Weather Station in Practice
Here we'll discuss basic stuff about importing my test data, running some tests on it, setting up your own data, displaying some graphs and setting up an email with prediction for the next few hours.
If you like to know more about how it works check out the next step "How does it work"
Importing my already gathered data
MongoDb comes with a cli command for importing data from json:
mongoimport -d weather -c weather_station --file sample_data/weather_station.json
This will import the file from sample data into "weather" database and "datapoints" collection
A warning here, if you use my gathered data, and combine it with your newly local data the accuracy might drop because of the small differences in hardware (sensors) and local weather patterns.
Gathering new data
One of the base station roles is to store incoming data from the weather station into the database for later processing. To start the process that listens to the serial port and stores in the database just run:
conda activate weather python serial_listener.py # every 10 minutes you should see data from the weather station coming in : [Sensor: type(temperature), value(14.3)] [Sensor: type(pressure), value(1056.0)] ...
Generating the prediction model
I'll presume you have imported my data or "ran the script for a few years" to gather your personialised data, so in this step we'll process the data to create a model used to predict future rain.
conda activate weather python train.py --days_behind 600 --test-file-percent 10 --datapoints-behind 8 --hour-granularity 6
* The first parameter --days_behind means how much data into the past should the script process. It's measured in days
* --test-file-percent means how much of the data should be considered for testing purpose, this is a regular step in a machine learning algorithm
* --hour-granularity basically means how many hours into the future we'll want the prediction
* --datapoints-behind this parameter will be discussed further into the next section
View some data charts with all the weather station sensors
Let's say for the last 10 days:
conda activate weather python graphs --days-behind 10
Predict if it will rain in the next period
We'll predict if it's going to rain and send an notification email
conda activate weather python predict.py --datapoints-behind 8 --hour-granularity 6 --from-addr a_gmail_address --from-password gmail_password --to-addr a_email_destination
Run a batch predict on the test data:
python predict_batch.py -f sample_data/test_data.csv
It's important to use the same parameters as in the train script above.
For the email notification to work log in into your gmail account and Turn Allow less secure apps to ON. Be aware that this makes it easier for others to gain access to your account.
You'll need two email addresses, a gmail address with the above option activated, and an other address where you'll receive your notification.
If you like to receive notifications every hour, put the script into crontab
To see how all of this is possible check the next step
How Does It Work
In this last step we'll discuss several aspects of the architecture of this project:
1. Project overview, we'll discuss the general architecture and technologies involved
2. Basic concepts of machine learning
3. How the data is prepared (the most important step)
4. How the actual neural network wrapper API works (Keras)
5. Future improvements
I'll try to give some code example here but bear in mind, that it's not 100% the code from the project. In the project it's self the code it's a bit more complicated with classes and a structure
1. Project overview, we'll discuss the general architecture and technologies involved
As we talked earlier, the project has two separate parts. The weather station it's self which only function is to collect and transmit data. And the base station where all the collection training and prediction will happen.
Advantages of the separation of the weather station and the base station:
- power requirements, if the weather station was able to process the data too, it would need substantial power, maybe large solar panels or a permanent power source
- portability, due to it's small size the weather station can collect data from a few hundred meters away and you can easily change it's place if needed
- scalability, you can increase the prediction accuracy by building more than one weather stations and spread them around a few hundred meters
- low cost, because it's a cheap device you can easily build another in case one is lost or stolen
The database choice. I've chosen mongoDb because it's nice features: schemaless, free and easy to use API
Every time sensor data is received the data is saved in the database, a the data entry looks something like this:
{ "_id" : "04_27_2017_06_17", "humidity" : 65, "date" : ISODate("2017-04-27T06:17:18Z"), "pressure" : 1007, "temperature" : 9, "rain" : 0, "light" : 15 }
The database stores data in BSON format (similar to JSON) so it's easy to read and easy to work with. I've aggregated the data under an identifier that contains the date formatted as string to the minutes, so the smallest grouping here is a minute.
The weather station (when working properly) will transmit a datapoint every 10 minutes. A datapoint is a collection of "date", "humidity", "pressure", "temperature", "rain" and "light" values.
The data processing and neural network technology choice
I've chosen python for the backend because many major inovations in neural networks are found in python. A growing community with a lot of Github repositories, tutorials blogs and books are here to help.
* For the data processing part i've used Pandas (https://pandas.pydata.org/). Pandas make working with data easy. You can load tables from CSV, Excel, python data structures and reorder them, drop columns, add columns, index by a column and many other transformations.
* For working with neural networks i've chosen Keras (https://keras.io/). Keras is a high level neural network wrapper over more lower level API's like Tensorflow and one can build a multi - layer neural network with a dozen lines of code or so. This is a great advantage because we can build something useful on the great work of other people. Well this is the basic stuff of programming, build on other smaller building blocks.
2. Basic concepts of machine learning
The scope of this tutorial is not to teach machine learning, but merely to outline one of it's possible use cases and how we can practically apply it to this use case.
Neural networks are data structures that resemble brain cells called neurons. Science discovered that a brain has special cells named neurons that communicate with other neurons by electrical impulses through "lines" called axons. If stimulated sufficiently (from many other neurons) the neurons will trigger an electric impulse further away in this "network" stimulating other neurons. This of course is a oversimplification of the process, but basically computer algorithms try to replicate this biological process.
In computer neural nets each neuron has a "trigger point" where if stimulated over that point it will propagate the stimulation forward, if not it won't. For this each simulated neuron will have a bias, and each axon a weight. After a random initialization of these values a process called "learning" starts this means in a loop an algorithm will do this steps:
- stimulate the input neurons
- propagate the signals through the network layers until the output neurons
- read the output neurons and compare the results with the desired results
- tweak the axons weights for a better result next time
- start again until the number of loops has been reached
If you like to know more details about this process you can check out this article: https://mattmazur.com/2015/03/17/a-step-by-step-ba.... There are also numerous books and tutorials out there.
One more thing, here we'll be using a supervised learning method. That means we'll teach the algorithm the inputs and the outputs also, so that given a new set of inputs it can predict the output.
3. How the data is prepared (the most important step)
In many machine learning and neural network problems data preparation is a very important part and it will cover:
- get the raw data
- data cleanup: this will mean removing orphan values, aberrations or other anomalies
- data grouping: taking many datapoints and transforming into an aggregated datapoint
- data enhancing: adding other aspects of the data derived from own data, or from external sources
- splitting the data in train and test data
- split each of the train and test data into inputs and outputs. Typically a problem will have many inputs and a few outputs
- rescale the data so it's between 0 and 1 (this will help the network removing high/low value biases)
Getting the raw data
In our case getting data for MongoDb in python is really easy. Given our datapoints collection just this lines of code will do
client = MongoClient(host, port).weather.datapoints cursor = client.find( {'$and' : [ {'date' : {'$gte' : start_date}}, {'date' : {'$lte' : end_date}} ]} ) data = list(cursor) ..
Data cleanup
The empty values in the dataframe are dropped
dataframe = dataframe.dropna()
Data grouping & data enhancing
This is a very important step, the many small datapoins will be grouped into intervals of 6 hours. For each group several metrics will be calculated on each of the sensors (humidity, rain, temperature, light, pressure)
- min value
- max value
- mean
- 70, 90, 30, 10 percentiles
- nr of times there has been a rise in a sensor
- nr of times there has been a fall in a sensor
- nr of times there has been steady values in a sensor
All of these things will give the network information for a datapoint, so for each of the 6 hours intervals these things will be known.
From a dataframe that looks like this:
_id date humidity light pressure rain temperature 04_27_2017_03_08 2017-04-27 03:08:36 67.0 0.0 1007.0 0.0 11.0 04_27_2017_03_19 2017-04-27 03:19:05 66.0 0.0 1007.0 0.0 11.0 04_27_2017_03_29 2017-04-27 03:29:34 66.0 0.0 1007.0 0.0 11.0
And the transformation will be:"
_id date humidity_10percentile humidity_30percentile humidity_70percentile humidity_90percentile humidity_avg ... temperature_avg temperature_fall temperature_max temperature_min temperature_rise temperature_steady ... 04_27_2017_0 2017-04-27 03:08:36 59.6 60.8 63.2 66.0 62.294118 ... 10.058824 2 11.0 9.0 1 14 04_27_2017_1 2017-04-27 06:06:50 40.3 42.0 60.0 62.0 50.735294 ... 14.647059 3 26.0 9.0 11 20 04_27_2017_2 2017-04-27 12:00:59 36.0 37.0 39.8 42.0 38.314286 ... 22.114286 1 24.0 20.0 5 29
After this a new column named "has_rain" will be added. This will be the output (our predicted variable). Has rain will be 0 or 1 depending if the rain average is above a threshold (0.1). With pandas it's as simple as:
dataframe.insert(loc=1, column='has_rain', value=numpy.where(dataframe['rain_avg'] > 0.1, 1, 0))
Data cleanup (again)
- we'll drop the date column because it's no use to us, and also remove datapoints where the minimum temperature is below 0 because our weather station it doesn't have a snow sensor, so we won't be able to measure if it snowed
dataframe = dataframe.drop(['date'], axis=1) dataframe = dataframe[dataframe['temperature_min'] >= 0]
Data enhancing
Because data in the past might influence our prediction of the rain, we need for each of the dataframe rows to add columns reference to the past rows. This is because each of the row will serve as a training point, and if we want the prediction of the rain to take into account previous datapoints that's exactly what we should do: add more columns for datapoints in the past ex:
_id has_rain humidity_10percentile humidity_30percentile humidity_70percentile humidity_90percentile ... temperature_steady_4 temperature_steady_5 temperature_steady_6 temperature_steady_7 temperature_steady_8 ... 04_27_2017_3 0 36.0 44.8 61.0 63.0 ... NaN NaN NaN NaN NaN 04_28_2017_0 0 68.0 70.0 74.0 75.0 ... 14.0 NaN NaN NaN NaN 04_28_2017_1 0 40.0 45.0 63.2 69.0 ... 20.0 14.0 NaN NaN NaN 04_28_2017_2 0 34.0 35.9 40.0 41.0 ... 29.0 20.0 14.0 NaN NaN 04_28_2017_3 0 36.1 40.6 52.0 54.0 ... 19.0 29.0 20.0 14.0 NaN 04_29_2017_0 0 52.0 54.0 56.0 58.0 ... 26.0 19.0 29.0 20.0 14.0 04_29_2017_1 0 39.4 43.2 54.6 57.0 ... 18.0 26.0 19.0 29.0 20.0 04_29_2017_2 1 41.0 42.0 44.2 47.0 ... 28.0 18.0 26.0 19.0 29.0
So you see that for every sensor let's say temperature the following rows will be added: "temperature_1" , "temperature_2" .. meaning temperature on the previous datapoint, temperature on the previous two datapoints etc. I've experimented with this and i found that a optimum number for our 6 hour groupings in 8. That means 8 datapoints in the past (48 hours). So our network learned the best from datapoins spanning 48 hours in the past.
Data cleanup (again)
As you see, the first few columns has "NaN" values because there is nothing in front of them so they should be removed because they are incomplete.
Also data about current datapoint should be dropped, the only exception is "has_rain". the idea is that the system should be able to predict "has_rain" without knowing anything but previous data.
Splitting the data in train and test data
This is very easy due to Sklearn package:
from sklearn.model_selection import train_test_split ... main_data, test_data = train_test_split(dataframe, test_size=percent_test_data) ...
This will split the data randomly into two different sets
Split each of the train and test data into inputs and outputs
Presuming that our "has_rain" interest column is located first
X = main_data.iloc[:, 1:].values y = main_data.iloc[:, 0].values
Rescale the data so it's between 0 and 1
Again fairly easy because of sklearn
from sklearn.preprocessing import StandardScaler from sklearn.externals import joblib .. scaler = StandardScaler() X = scaler.fit_transform(X) ... # of course we should be careful to save the scaled model for later reuse joblib.dump(scaler, 'model_file_name.save')
4. How the actual neural network wrapper API works (Keras)
Building a multi layer neural network with Keras is very easy:
from keras.models import Sequential from keras.layers import Dense from keras.layers import Dropout ... input_dimensions = X.shape[1] optimizer = 'rmsprop' dropout = 0.05 model = Sequential() inner_nodes = int(input_dimensions / 2) model.add(Dense(inner_nodes, kernel_initializer='uniform', activation='relu', input_dim=input_dimensions)) model.add(Dropout(rate=dropout)) model.add(Dense(inner_nodes, kernel_initializer='uniform', activation='relu')) model.add(Dropout(rate=dropout)) model.add(Dense(1, kernel_initializer='uniform', activation='sigmoid')) model.compile(optimizer=optimizer, loss='mean_absolute_error', metrics=['accuracy']) model.fit(X, y, batch_size=1, epochs=50) ... # save the model for later use classifier.save('file_model_name')
So what does this code mean? Here we're building a sequential model , that means sequentially all the layers will be evaluated.
a) we declare the input layer (Dense), here all the inputs from our dataset will be initializedm so the "input_dim" parameter must be equal to the row length
b) a Dropout layer is added. To understand the Dropout first we must understand what "overfitting" means: it's a state in which the network has learned too much particularities for a specific dataset and will perform badly when confronted to a new dataset. The dropout layer will disconnect randomly neurons at each iteration so the network won't overfit.
c) another layer of Dense is added
d) another Dropout
e) the last layer is added with one output dimension (it will predict only yes/no)
f) the model is "fitted" that means the learning process will begin, and the model will learn
Other parameters here:
- activation functions (sigmoid, relu). This are functions that dictate when the neuron will transmit it's impulse further in the network. There are many, but sigmoid and relu are the most common. Check out this link for more details: https://towardsdatascience.com/activation-function...
- kernel_initializer function (uniform). This means that all the weights are initialized with random uniform values
- loss function (mean_absolute_error). This function measures the error comparing the network predicted result versus the ground truth. There are many alternatives: https://keras.io/losses/
- metrics function (accuracy). It measures the performance of the model
- optimiser functions (rmsprop). It optimizes how the model learn through backpropagation.
- batch_size. Number of datapoints to take once by Keras before applying optimizer function
- epochs: how many times the process it's started from 0 (to learn better)
There is no best configuration for any network or dataset, all these parameters can an should be tuned for optimal performance and will make a big difference in prediction success.
5. Future improvements
Let's start from the weather station, i can see a lot of improving work to be done here:
- add a wind speed / direction sensor. This could be a very important sensor that i'm missing in my model
- experiment with UV rays, gas and particle sensors
- add at least two stations in the zone for better data (make some better averages)
- collect a few more years of data, i've experimented with just a year and a half
Some processing improvements:
- try to incorporate data from other sources into the model. You can start to import wind speed data and combine with the local station data for a better model. This website offers historical data: https://www.wunderground.com/history/
- optimize the Keras model better by adjusting: layers, nr of neurons in layers, dropout percents, metrics functions, optimiser functions, loss functions, batch size, learning epochs
- try other model architectures, for example i've experimented with LSTM (long short term memory) but it gived slightly poorer results)
To try different parameters on the learning model you can use
python train.py --days_behind 600 --test-file-percent 10 --datapoints-behind 6 --hour-granularity 6 --grid-search
This will search through different "batch_size", "epoch", "optimizer" and "dropout" values, evaluate all and print out the best combination for your data.
If you have some feedback on my work please share it, thanks for staying till the end of the tutorial!
Bonus: Using an Official Weather Dataset
I was wondering if i can get better results with a more reliable weather station, so i've searched a bit, and i've came across "Darksky API" (https://darksky.net/dev), this is a great tool that provides current and historical weather data with many more sensor data:
- temperature
- humidity
- pressure
- wind speed
- wind gust
- ub index
- visibility
So this beeing data from an official weather station, and having more parameters i thought it should perform better so i've gave it a try. To replicate my findings:
1. Download the data from darsky or import my MongoDb collection:
a) Download
- to download your self, first create an account in darsky and get the API key
- replace the API key in download_import/config.py
- also in the config replace the geographic coordonates for the location you want to predict the rain
- in a console activate "weather" anaconda environment and run:
python download_import/darksky.py -d 1000
- the free version of the API is limited to 1000 requests per day so if you want more data you need to wait for a longer time
b) Import my downloaded data for Bucharest city
- in a console run
mongoimport -d weather -c darksky --file sample_data/darksky.json
2. When you train the model specify that it should run on "darksy" dataset
python train.py -d 2000 -p 20 -dp 4 -hg 6 --data-source darksky
3. To see the results run predict batch script as before
python predict_batch.py -f sample_data/test_data.csv
You'll see that the overall prediction percent has gone from about 80% to 90%. Also the prediction accuracy when accounting only rainy days has gone up.
So yes the dataset it really matters