*before we do this lesson, you will want to open a terminal on your desktop and pip install Requests, Pillow, Pandas, NumPy, Scikit-Learn, Tensorflow, and matplotlib.
Learning Objectives:
Select approproate libraries or existing code segments to use in creating new programs
What are Libraries?
In Python, a library is a collection of pre-written code, functions, and modules that extend the language’s capabilities. These libraries are designed to be reused by developers to perform common tasks, rather than having to write the code from scratch. Libraries are essential for simplifying and accelerating the development process, as they provide a wide range of tools and functions for various purposes.
Here are some key points about Python libraries:
Modules: Libraries in Python consist of modules, which are individual Python files containing functions, classes, and variables related to a specific set of tasks or a particular domain. You can import these modules into your own Python code to access their functionality.
Standard Library: Python comes with a comprehensive standard library that includes modules for various tasks, such as working with files, networking, data processing, and more. These modules are readily available and do not require installation.
Third-party Libraries: In addition to the standard library, there is a vast ecosystem of third-party libraries created by the Python community. These libraries cover a wide range of domains, including web development, data analysis, machine learning, game development, and more. Some popular third-party libraries include NumPy, Pandas, Matplotlib, TensorFlow, Django, Flask, and many others.
How Do We Get Libraries into Our Code and Working?</strong>
To get libraries into our code, we use the import statement followed by the library we want to import. Lets start simply:
#In this code cell, we are importing the math library which allows us to do math operations,
#and the random library which lets us take pseudorandom numbers and choices.
importmathimportrandom#We use the libraries by first calling them by their name, then using one of their methods.
#For example:
num=64print(math.sqrt(num))numList=[1,2,3,4,5,6]print(random.choice(numList))#Here, 'math' and 'random' are the names of the libraries, and 'sqrt' and 'choice' are the names of the methods.
We can also import parts of libraries by adding a "from" in front of our import.
Now, we don't have to use math. in front of sqrt, and can just use the function by itself. We can also import *, or all, which makes it so that everything is imported. Here, we don't have to use random in front of choice, even though we didn't import choice specifically.
Popcorn Hack #1
Import your own library from a list of provided libraries, and use one of its methods. This can be something very bare bones, such as printing the time, getting a random number in a list, or doing something after sleeping a certain amount of time
```python
#math library module examples: sqrt(num), square(num), cube(num), factorial(num)
#random library module examples: choice(list), randrange(lowest, highest, step[numbers chosen in multiples of {step}])
#datetime library module examples: datetime.now()
#sleep library module examples: sleep(milliseconds)
from math import factorial
num = 5
print (factorial(num))
```
120
Documentation</strong>
Documentation in Python libraries refers to the written information and explanations provided to help users understand how to use the library, its classes, functions, and modules. It serves as a comprehensive guide that documents the library's functionality, usage, and often includes code examples. Documentation is typically created by the library developers and is an essential component of a well-maintained library.
Examples of Documentation: An introductory section explaining the purpose of the library, a section on how to install the library, basic usage examples, etc.
```python
calcAverage(grades)
'''
You know the name of the procedure and the perameters, but...
You probably wouldn't be able to use this procedure
with confidence because you don't know its function
exactly (maybe you can guess that it finds the average,
but you wouldn't know if it uses mean, mode, or median to
find the average). You would also need more information on the
perameters.
'''
```
Libraries and APIs</strong>
- A file that contains procedures that cane be used in a program is called a library
- An Application Program Interface (API) provides specifications for how procedures in a library behave and can be used.
APIs define the methods and functions that are available for developers to interact with a library. They specify how to make requests, provide inputs, and receive outputs, creating a clear and consistent way to use library features.
Which libraries will be very important to us?
Requests - Simplifies working with HTTP servers, including 'request'-ing data from them, and recieving it
Pillow - Simplifies image processing
Pandas - Simplifies data analysis & manipulation
Numpy - Vastly quickens functionality of arrays up to 50 times faster than regular python list
Scikit-Learn - Implements machine learning models and statistical modelling
Tensorflow - Data automation, model tracking, performance monitoring, and model retraining
Matplotlib - Creates static, animated, and interactive visualizations in Python
DON'T FORGET TO DOWNLOAD ALL OF THESE (pip install "library")
Popcorn Hack #2
Using the requests library and the ? module (since we should already be using this in our backend) GET a request from the api at "https://api.github.com"
```python
import requests
#GET a request using the requests library. Remember to put your api link in quotes! If you get something along the lines of response [200] then you succeeded
x = requests.get('http://127.0.0.1:4200/student2/')
print(x.text)
```
<!DOCTYPE html>
Aidan’s Blog | csp BlogsSkip to the content.
My name is Aidan Delgado and I am a junior at Del Norte
I have two cats
I run track
I like to eat
My freeform drawing:
Picture collage!
Challenges I Faced:
At first, my bundle install would not work because I did not have jupyter properly installed. So, I had to reinstall all my updates again and bundle install finally worked.
When changing my theme, my config.yml page would always have a problem whenever I tried to edit the theme on there. My config.yml page would always have an error in it no matter what I changed, even when I copied and pasted the default code off of the teacher github page. I tried over and over again to find the problem, when I realized I could copy and paste one of my classmate’s codes that worked for them. I did this and my theme changed to the one I wanted.
## Scikit-Learn and Numpy
This code uses NumPy to create an array, and Scikit-Learn to analyze the data. It creates a linear regression which describes the relationship between the x and y arrays which reperesent independent and dependent variables. In simpler terms, it is creating a line of best fit between the two data sets, just like how you would in something like desmos.
>
```python
import numpy as np
import numpy as np
from sklearn.linear_model import LinearRegression
from sklearn.model_selection import train_test_split
from sklearn.metrics import mean_squared_error
# Generate some example data
X = np.array([1, 2, 3, 4, 5]).reshape(-1, 1) # Feature (independent variable)
y = np.array([2, 4, 5, 4, 5]) # Target (dependent variable)
# Split the data into training and testing sets
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)
# Create a Linear Regression model
model = LinearRegression()
# Fit the model to the training data
model.fit(X_train, y_train)
# Make predictions on the test data
y_pred = model.predict(X_test)
# Evaluate the model by calculating the Mean Squared Error
mse = mean_squared_error(y_test, y_pred)
# Print the model coefficients and MSE, model coefficient is the slope of the linear regression line, MSE is how well the model is performing, the closer it is to 0 the better
print("Model Coefficients:", model.coef_)
print("Mean Squared Error:", mse)
intercept = model.intercept_
slope = model.coef_[0]
print(f"Linear Regression Equation: y = {intercept} + {slope} * X")
```
## Request
- The requests module allows you to send HTTP requests using Python.
- In order to download requests, you would have to type pip install requests in your terminal.
## Syntax
- requests.methodname(params)
```python
import requests
x = requests.get('http://127.0.0.1:9008/')
print(x.text)
#not functional code, example of syntax
```
```python
import requests
# Replace this URL with the website you want to request
url = 'https://www.example.com'
# Send a GET request to the URL
response = requests.get(url)
# Check if the request was successful (status code 200)
if response.status_code == 200:
# Print the content of the response (the HTML content of the webpage)
print(response.text)
else:
# If the request was not successful, print an error message
print(f"Failed to retrieve the page. Status code: {response.status_code}")
```
## Pillow
- Pillow is a imaging library that provides easy-to-use methods to include, change, save different image formats.
- To dowload pillow onto your computer, you would enter the command pip install Pillow into your terminal.
```python
from PIL import Image, ImageDraw, ImageFont
# Create a new blank image
width, height = 400, 200
image = Image.new("RGB", (width, height), "white")
# Create an ImageDraw object
draw = ImageDraw.Draw(image)
# Draw a red line from (50, 50) to (350, 150)
line_color = (255, 0, 0) # Red color
draw.line((50, 50, 350, 150), fill=line_color, width=3)
# Add text to the image
text = "This was created using Pillow!"
text_color = (0, 0, 0) # Black color
font_size = 20
font = ImageFont.load_default() # Use a default font
text_position = (50, 160)
draw.text(text_position, text, fill=text_color, font=font)
# Save or display the image
image.show()
image.show()
#This opens the image in your default image viewer and when you stop the code, it will return an error, but don't worry about that
```
```python
from PIL import Image
# Open an image
original_image = Image.open('image.png') #replace image.png with valid image
# Display information about the image
width, height = original_image.size
format = original_image.format
print(f"Original Image Size: {width}x{height}")
print(f"Original Image Format: {format}")
# Resize the image to a new size
new_size = (width // 2, height // 2) # Reduce the size by half
resized_image = original_image.resize(new_size)
# Save the resized image
resized_image.save('resized_image.jpg')
# Display information about the resized image
resized_width, resized_height = resized_image.size
print(f"Resized Image Size: {resized_width}x{resized_height}")
# Show both the original and resized images
original_image.show()
resized_image.show()
```
---------------------------------------------------------------------------
FileNotFoundError Traceback (most recent call last)
c:\Users\matth\librarylesson\lessons\2023-10-26-Libraries.ipynb Cell 26 line 4
1 from PIL import Image
3 # Open an image
----> 4 original_image = Image.open('image.png')
6 # Display information about the image
7 width, height = original_image.size
File c:\Users\matth\AppData\Local\Programs\Python\Python311\Lib\site-packages\PIL\Image.py:3243, in open(fp, mode, formats)
3240 filename = fp
3242 if filename:
-> 3243 fp = builtins.open(filename, "rb")
3244 exclusive_fp = True
3246 try:
FileNotFoundError: [Errno 2] No such file or directory: 'image.png'
## Pandas
This code utilizes pandas in the DataFrame form to organize the data in to a table with the categories on the horizontal axis and their values on the vertical. Pandas creates a way for the user to organize data in a much simpler form and in different styles depending on what the user wants
```python
#This code utilizes pandas which is a way for you as a user, to create data tables that are much more organized
#imports pandas so it's able to be used
import pandas as pd
#data is created and will be sorted from left to right into top to bottom
data = {'Name': ['Matthew', 'Lindsay', 'Josh', 'Ethan'],
'Grade': [97, 92, 90, 80]}
#defines a variable and utilizes pandas by using a DataFrame for the data
df = pd.DataFrame(data)
#tbere are other forms other than DataFrame, those are Series (single Column), Panel (3D), Multindex (multiple levels of index), and Categorical (categories),
#increase the side count by one
df.index += 1
print(df)
```
Name Grade
1 Matthew 97
2 Lindsay 92
3 Josh 90
4 Ethan 80
Another example code, this time utilizing both numpy and pandas
```python
import pandas as pd
import numpy as np
# Sample data
data = {
'Grade': ['A', 'B', 'A', 'C', 'B', 'C', 'A', 'B', 'A'],
'Percent': [94, 82, 91, 76, 89, 79, 92, 87, 99]
}
# Create a Pandas DataFrame
df = pd.DataFrame(data)
# Calculate the mean of each Grade using NumPy
means = df.groupby('Grade')['Percent'].mean().reset_index()
# Organize the results into a new data table
result = pd.DataFrame({'Grade': ['A', 'B', 'C'], 'Mean Grade': means['Percent']})
result.index += 1
# Display the result
print(result)
```
Grade Mean Grade
1 A 94.0
2 B 86.0
3 C 77.5
## TensorFlow
The provided code demonstrates a basic example of linear regression using TensorFlow and Keras. It begins by importing the necessary libraries, TensorFlow and NumPy. It then generates a synthetic dataset with 1000 samples, where the input features are random, and the target values are computed as a linear combination of the input features with added noise. A data pipeline is set up using TensorFlow, which includes shuffling and batching the data for efficient processing. A simple linear regression model is defined using Keras, consisting of one dense layer. The model is compiled with the Adam optimizer and mean squared error as the loss function. It is then trained on the synthetic data for ten epochs. After training, the model is used to make predictions on new data points, and the predictions are printed to the console. This code provides a basic illustration of how to perform a simple machine learning task with TensorFlow, from data generation to model training and prediction.
```python
import tensorflow as tf
import numpy as np
# Create a synthetic dataset
num_samples = 1000
input_data = np.random.rand(num_samples, 2)
target_data = input_data[:, 0] * 2 + input_data[:, 1] * 3 + np.random.randn(num_samples)
# Define a data pipeline using TensorFlow
dataset = tf.data.Dataset.from_tensor_slices((input_data, target_data))
dataset = dataset.shuffle(buffer_size=num_samples)
dataset = dataset.batch(32)
dataset = dataset.prefetch(buffer_size=tf.data.AUTOTUNE)
# Create a simple linear regression model using Keras
model = tf.keras.Sequential([
tf.keras.layers.Dense(1, input_shape=(2,))
])
model.compile(optimizer='adam', loss='mean_squared_error')
# Train the model on the synthetic data
model.fit(dataset, epochs=10)
# Generate predictions
new_data = np.array([[0.5, 0.7], [0.3, 0.2]])
predictions = model.predict(new_data)
print("Predictions:", predictions)
```
Epoch 1/10
32/32 [==============================] - 0s 2ms/step - loss: 3.8766
Epoch 2/10
32/32 [==============================] - 0s 2ms/step - loss: 3.6717
Epoch 3/10
32/32 [==============================] - 0s 2ms/step - loss: 3.4812
Epoch 4/10
32/32 [==============================] - 0s 2ms/step - loss: 3.3002
Epoch 5/10
32/32 [==============================] - 0s 2ms/step - loss: 3.1319
Epoch 6/10
32/32 [==============================] - 0s 1ms/step - loss: 2.9735
Epoch 7/10
32/32 [==============================] - 0s 2ms/step - loss: 2.8254
Epoch 8/10
32/32 [==============================] - 0s 2ms/step - loss: 2.6876
Epoch 9/10
32/32 [==============================] - 0s 2ms/step - loss: 2.5581
Epoch 10/10
32/32 [==============================] - 0s 2ms/step - loss: 2.4363
1/1 [==============================] - 0s 85ms/step
Predictions: [[1.6355244]
[0.9337466]]
## Matplotlib
The provided Python code demonstrates the basic usage of Matplotlib, a popular library for creating data visualizations. In this example, we start by importing the Matplotlib's pyplot module, often aliased as plt. We define some sample data as lists for the X and Y values. Then, we create a figure and an axis object using plt.subplots(). Next, we plot the data points on the graph with ax.plot(x, y) and set a label for the line. We also add labels for the X and Y axes and set a title for the plot. To provide context for the plot, we include a legend with the label we set earlier. Finally, plt.show() is called to display the graph. When you run this code, it will generate a simple line plot displaying the data points with appropriate labels, a title, and a legend, making it a clear and informative visualization.
```python
import matplotlib.pyplot as plt
# Sample data
x = [1, 2, 3, 4, 5]
y = [2, 4, 6, 8, 10]
# Create a figure and axis
fig, ax = plt.subplots()
# Plot the data
ax.plot(x, y, label='Linear Line')
# Set labels and title
ax.set_xlabel('X-axis')
ax.set_ylabel('Y-axis')
ax.set_title('Simple Line Plot')
# Add a legend
ax.legend()
# Display the plot
plt.show()
```
![png](output_33_0.png)
Homework Hack
1) Create a code that makes a data table which organizes the average values(mean) from a data set the has atleast 5 values per category and using 2 libraries, ex: </br>
2) Create a Python script that downloads images from a website using the requests library, processes them with the Pillow library, and then performs data analysis with the Pandas library.
```python
#homework 1
import pandas as pd
import numpy as np
data = {
'Category': ['A', 'B', 'A', 'C', 'B'],
'Value': [10, 15, 12, 20, 18]
}
df = pd.DataFrame(data)
mean = df.groupby('Category')['Value'].mean().reset_index()
mean.columns = ['Category', 'Mean']
print(mean)
```
Category Mean
0 A 11.0
1 B 16.5
2 C 20.0
```python
#homework 2
import os
import requests
from PIL import Image
import pandas as pd
# Function to download an image from a URL
def download_image(url, save_path):
response = requests.get(url)
if response.status_code == 200:
with open(save_path, 'wb') as file:
file.write(response.content)
# Function to process an image (e.g., resize it)
def process_image(image_path, output_path, size=(100, 100)):
image = Image.open(image_path)
image = image.resize(size)
image.save(output_path)
# Directory to save downloaded images
output_directory = 'downloaded_images'
os.makedirs(output_directory, exist_ok=True)
image_data = []
for i, url in enumerate(image_urls):
image_filename = f'image_{i}.jpg'
image_path = os.path.join(output_directory, image_filename)
process_path = os.path.join(output_directory, f'processed_{image_filename}')
image_data.append({
'Image URL': url,
'Image Path': image_path,
'Processed Image Path': process_path,
})
df = pd.DataFrame(image_data)
print(df.head())
```
https://www.google.com/search?sca_esv=576910264&rlz=1C1VDKB_enUS1071US1071&sxsrf=AM9HkKkVKA1Zn1w0qR61y-JWCyAURCI8zw:1698354078544&q=1000+yard+stare&tbm=isch&source=lnms&sa=X&ved=2ahUKEwi-s7f8zZSCAxXYHEQIHTsdCeoQ0pQJegQIFBAB&biw=1366&bih=611&dpr=1#imgrc=jNdwnZupWZw_fM