diff --git a/ML Projects/Lane Line Detection/Readme.md b/ML Projects/Lane Line Detection/Readme.md new file mode 100644 index 000000000..7e9ca9e4c --- /dev/null +++ b/ML Projects/Lane Line Detection/Readme.md @@ -0,0 +1 @@ +Lane Line detection using open cv diff --git a/ML Projects/Lane Line Detection/gui.py b/ML Projects/Lane Line Detection/gui.py new file mode 100644 index 000000000..045d87f5e --- /dev/null +++ b/ML Projects/Lane Line Detection/gui.py @@ -0,0 +1,74 @@ +import tkinter as tk +from tkinter import * +import cv2 +from PIL import Image, ImageTk +import os +import numpy as np + + +global last_frame1 #creating global variable +last_frame1 = np.zeros((480, 640, 3), dtype=np.uint8) +global last_frame2 #creating global variable +last_frame2 = np.zeros((480, 640, 3), dtype=np.uint8) +global cap1 +global cap2 +cap1 = cv2.VideoCapture("./test2.mp4") +cap2 = cv2.VideoCapture("./test2.mp4") + +def show_vid(): + if not cap1.isOpened(): + print("cant open the camera1") + flag1, frame1 = cap1.read() + frame1 = cv2.resize(frame1,(600,500)) + if flag1 is None: + print ("Major error!") + elif flag1: + global last_frame1 + last_frame1 = frame1.copy() + pic = cv2.cvtColor(last_frame1, cv2.COLOR_BGR2RGB) + img = Image.fromarray(pic) + imgtk = ImageTk.PhotoImage(image=img) + lmain.imgtk = imgtk + lmain.configure(image=imgtk) + lmain.after(10, show_vid) + + +def show_vid2(): + if not cap2.isOpened(): + print("cant open the camera2") + flag2, frame2 = cap2.read() + frame2 = cv2.resize(frame2,(600,500)) + if flag2 is None: + print ("Major error2!") + elif flag2: + global last_frame2 + last_frame2 = frame2.copy() + pic2 = cv2.cvtColor(last_frame2, cv2.COLOR_BGR2RGB) + img2 = Image.fromarray(pic2) + img2tk = ImageTk.PhotoImage(image=img2) + lmain2.img2tk = img2tk + lmain2.configure(image=img2tk) + lmain2.after(10, show_vid2) + +if __name__ == '__main__': + root=tk.Tk() + img = ImageTk.PhotoImage(Image.open("logo.png")) + heading = Label(root,image=img, text="Lane-Line Detection") + # heading.configure(background='#CDCDCD',foreground='#364156') + heading.pack() + heading2=Label(root,text="Lane-Line Detection",pady=20, font=('arial',45,'bold')) + heading2.configure(foreground='#364156') + heading2.pack() + lmain = tk.Label(master=root) + lmain2 = tk.Label(master=root) + + lmain.pack(side = LEFT) + lmain2.pack(side = RIGHT) + root.title("Lane-line detection") + root.geometry("1250x900+100+10") + + exitbutton = Button(root, text='Quit',fg="red",command= root.destroy).pack(side = BOTTOM,) + show_vid() + show_vid2() + root.mainloop() + cap.release() \ No newline at end of file diff --git a/ML Projects/Lane Line Detection/logo.png b/ML Projects/Lane Line Detection/logo.png new file mode 100644 index 000000000..bf8cf96c1 Binary files /dev/null and b/ML Projects/Lane Line Detection/logo.png differ diff --git a/ML Projects/Lane Line Detection/main.py b/ML Projects/Lane Line Detection/main.py new file mode 100644 index 000000000..3de8368ba --- /dev/null +++ b/ML Projects/Lane Line Detection/main.py @@ -0,0 +1,146 @@ +import matplotlib.pyplot as plt + +import numpy as np +import cv2 +import os +import matplotlib.image as mpimg +from moviepy.editor import VideoFileClip +import math + +def interested_region(img, vertices): + if len(img.shape) > 2: + mask_color_ignore = (255,) * img.shape[2] + else: + mask_color_ignore = 255 + + cv2.fillPoly(np.zeros_like(img), vertices, mask_color_ignore) + return cv2.bitwise_and(img, np.zeros_like(img)) + +def hough_lines(img, rho, theta, threshold, min_line_len, max_line_gap): + lines = cv2.HoughLinesP(img, rho, theta, threshold, np.array([]), minLineLength=min_line_len, maxLineGap=max_line_gap) + line_img = np.zeros((img.shape[0], img.shape[1], 3), dtype=np.uint8) + # print(lines) + lines_drawn(line_img,lines) + return line_img + +def lines_drawn(img, lines, color=[255, 0, 0], thickness=6): + global cache + global first_frame + slope_l, slope_r = [],[] + lane_l,lane_r = [],[] + + α =0.2 + + + for line in lines: + for x1,y1,x2,y2 in line: + slope = (y2-y1)/(x2-x1) + if slope > 0.4: + slope_r.append(slope) + lane_r.append(line) + elif slope < -0.4: + slope_l.append(slope) + lane_l.append(line) + #2 + img.shape[0] = min(y1,y2,img.shape[0]) + + # to prevent errors in challenge video from dividing by zero + if((len(lane_l) == 0) or (len(lane_r) == 0)): + print ('no lane detected') + return 1 + + #3 + slope_mean_l = np.mean(slope_l,axis =0) + slope_mean_r = np.mean(slope_r,axis =0) + mean_l = np.mean(np.array(lane_l),axis=0) + mean_r = np.mean(np.array(lane_r),axis=0) + + if ((slope_mean_r == 0) or (slope_mean_l == 0 )): + print('dividing by zero') + return 1 + + x1_l = int((img.shape[0] - mean_l[0][1] - (slope_mean_l * mean_l[0][0]))/slope_mean_l) + x2_l = int((img.shape[0] - mean_l[0][1] - (slope_mean_l * mean_l[0][0]))/slope_mean_l) + x1_r = int((img.shape[0] - mean_r[0][1] - (slope_mean_r * mean_r[0][0]))/slope_mean_r) + x2_r = int((img.shape[0] - mean_r[0][1] - (slope_mean_r * mean_r[0][0]))/slope_mean_r) + + #6 + if x1_l > x1_r: + x1_l = int((x1_l+x1_r)/2) + x1_r = x1_l + y1_l = int((slope_mean_l * x1_l ) + mean_l[0][1] - (slope_mean_l * mean_l[0][0])) + y1_r = int((slope_mean_r * x1_r ) + mean_r[0][1] - (slope_mean_r * mean_r[0][0])) + y2_l = int((slope_mean_l * x2_l ) + mean_l[0][1] - (slope_mean_l * mean_l[0][0])) + y2_r = int((slope_mean_r * x2_r ) + mean_r[0][1] - (slope_mean_r * mean_r[0][0])) + else: + y1_l = img.shape[0] + y2_l = img.shape[0] + y1_r = img.shape[0] + y2_r = img.shape[0] + + present_frame = np.array([x1_l,y1_l,x2_l,y2_l,x1_r,y1_r,x2_r,y2_r],dtype ="float32") + + if first_frame == 1: + next_frame = present_frame + first_frame = 0 + else : + prev_frame = cache + next_frame = (1-α)*prev_frame+α*present_frame + + cv2.line(img, (int(next_frame[0]), int(next_frame[1])), (int(next_frame[2]),int(next_frame[3])), color, thickness) + cv2.line(img, (int(next_frame[4]), int(next_frame[5])), (int(next_frame[6]),int(next_frame[7])), color, thickness) + + cache = next_frame + + +# def hough_lines(img, rho, theta, threshold, min_line_len, max_line_gap): +# lines = cv2.HoughLinesP(img, rho, theta, threshold, np.array([]), minLineLength=min_line_len, maxLineGap=max_line_gap) +# line_img = np.zeros((img.shape[0], img.shape[1], 3), dtype=np.uint8) +# lines_drawn(line_img,lines) +# return line_img + +def weighted_img(img, initial_img, α=0.8, β=1., λ=0.): + return cv2.addWeighted(initial_img, α, img, β, λ) + + +def process_image(image): + + global first_frame + + gray_image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) + img_hsv = cv2.cvtColor(image, cv2.COLOR_RGB2HSV) + + + lower_yellow = np.array([20, 100, 100], dtype = "uint8") + upper_yellow = np.array([30, 255, 255], dtype="uint8") + + mask_yellow = cv2.inRange(img_hsv, lower_yellow, upper_yellow) + mask_white = cv2.inRange(gray_image, 200, 255) + mask_yw = cv2.bitwise_or(mask_white, mask_yellow) + mask_yw_image = cv2.bitwise_and(gray_image, mask_yw) + + gauss_gray= cv2.GaussianBlur(mask_yw_image, (5, 5), 0) + + + canny_edges=cv2.Canny(gauss_gray, 50, 150) + + imshape = image.shape + lower_left = [imshape[1]/9,imshape[0]] + lower_right = [imshape[1]-imshape[1]/9,imshape[0]] + top_left = [imshape[1]/2-imshape[1]/8,imshape[0]/2+imshape[0]/10] + top_right = [imshape[1]/2+imshape[1]/8,imshape[0]/2+imshape[0]/10] + vertices = [np.array([lower_left,top_left,top_right,lower_right],dtype=np.int32)] + roi_image = interested_region(canny_edges, vertices) + + theta = np.pi/180 + + line_image = hough_lines(roi_image, 4, theta, 30, 100, 180) + result = weighted_img(line_image, image, α=0.8, β=1., λ=0.) + return result + +if __name__ == "__main__": + first_frame = 1 + white_output = './output.mp4' + clip1 = VideoFileClip(filename='test2.mp4') + white_clip = clip1.fl_image(process_image) + white_clip.write_videofile(white_output, audio=False) \ No newline at end of file