#commonly-used helper functions for the other scripts #By Drew Murray murraydr@msu.edu import pandas as pd import copy as cp import smtplib import os.path from datetime import datetime, timedelta import csv #libraries for google forms from apiclient import discovery from httplib2 import Http from oauth2client import client, file, tools from google.oauth2.credentials import Credentials from google_auth_oauthlib.flow import InstalledAppFlow from googleapiclient.discovery import build from google.auth.transport.requests import Request sectionDatabaseFilename="sectionsDatabase.csv" staffDatabaseFilename="staffDatabase.csv" #Converts full name to nedID based on the staff database #Uses a mildly 'fuzzy' search where the searched-for string must only occur *somewhere* in the actual full name def nameToID(searchName,getAllMatches=False): if searchName == "": return "" stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) try: ret = stfD[stfD['Name'].str.contains(searchName.strip("*"),case=False,regex=False)]["NetID"].values except IndexError: print ('ERROR: name "'+searchName+'" does not appear in the staff database') ret = -1 if getAllMatches: return ret else: return ret[0] #Converts netID to full name for GUI human-readability def IDToName(searchID): if searchID=="": return "" stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) name = stfD.loc[stfD['NetID']==searchID].values[0][0] if stfD.loc[stfD['NetID']==searchID].values[0][3]=="1":#if returning staff name+="*"#append asterisk to name return name #Tests if two requests overlap i.e. have at least one day/time combo in common def isRequestOverlap(req1,req2): print(req1) print(req2) sections1=req1[2].split(';') sections2=req2[2].split(';') dates1=req1[3].split(';') dates2=req2[3].split(';') overlap=False for s1 in sections1: for s2 in sections2: for d1 in dates1: for d2 in dates2: if s1==s2 and d1==d2: overlap=True return overlap #Get the time that a section takes place during def getTimeFromSection(section): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) return(secD.loc[secD['Section']==section]['Time'].iloc[0]) #Get the location that a section takes place in def getLocationFromSection(section): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) return(secD.loc[secD['Section']==section]['Location'].iloc[0]) #Get all dates that a section takes place in def getDatesFromSection(sec): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) dates = list(secD.loc[secD["Section"]==sec]["Date"]) return dates #Get the current and original data associated with a given section def getAllDataForSection(sec): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) secDOrig = pd.read_csv(sectionDatabaseFilename[:-4]+"_Orig.csv",dtype=str,index_col=False) ret=secD.loc[secD["Section"]==sec].dropna(axis=1,how='all').drop('Section',axis=1).fillna('') ret2=secDOrig.loc[secDOrig["Section"]==sec].dropna(axis=1,how='all').drop('Section',axis=1).fillna('') return ret,ret2 #Tests if a netID is currently assigned to a given section on a given date def isAssigned(netID,period,date): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) ret= False filt = (secD.loc[(secD['Section']==period) & (secD['Date']==date)].apply(lambda r: r.astype('string').str.contains(netID).any(), axis=1)) if filt.any(): #if the requestor is among the names in that section-date ret=True return ret #Test if a ULA is assigned to ANY section at the given date-time, return the section if True, else return False def isBusy(netID,time,date): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) filt = secD.loc[(secD['Time']==time) & (secD['Date']==date)].apply(lambda r: r.astype('string').str.contains(netID).any(),axis=1) if filt.any(): return secD.loc[(secD['Time']==time) & (secD['Date']==date) & (secD.isin([netID]).any(axis=1))]['Section'].to_string(index=False) else: return False #Overwrite a netID in one row of the section database with another def reassign(date,period,oldID,newID): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) secD.loc[(secD['Section']==period) & (secD['Date']==date)]=secD.loc[(secD['Section']==period) & (secD['Date']==date)].replace(oldID,newID) secD.to_csv(sectionDatabaseFilename,index=False,index_label=False) #Get the strings of all section titles currently in the database (dropping duplicates) def getAllSectionTitles(): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) sections=list(secD.drop_duplicates(subset=["Section"])["Section"].values) return sections #Get the strings of all names currently in the staff database def getAllNames(): stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) return list(stfD["Name"].values) #Add or subtract from the count of a given person's sub request history in one category (e.g. rejections) def incrementSubCount(netID,category,amount=1): stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) if category==0 or category=="APP": columnName="Approved Substitutions" elif category==1 or category=="ACC": columnName="Accepted Substitutions" elif category==2 or category=="REJ": columnName="Rejected Substitutions" elif category==3 or category=="CAN": columnName="Cancelled Substitutions" elif category==4 or category=="FUL": columnName="Fulfilled Substitutions" else: print("ERROR: Invalid category to incrementSubCount") return stfD.loc[stfD['NetID']==netID,columnName]=str(float(stfD.loc[stfD['NetID']==netID,columnName])+amount) stfD.to_csv(staffDatabaseFilename,index=False,index_label=False) #Get an array of the counts of each category of sub history (Approved, Accepted, Rejected, Cancelled, Fulfilled) for a given person def getSubCount(netID): stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) APP = stfD.loc[(stfD["NetID"]==netID)]["Approved Substitutions"].values[0] ACC = stfD.loc[(stfD["NetID"]==netID)]["Accepted Substitutions"].values[0] REJ = stfD.loc[(stfD["NetID"]==netID)]["Rejected Substitutions"].values[0] CAN = stfD.loc[(stfD["NetID"]==netID)]["Cancelled Substitutions"].values[0] FUL = stfD.loc[(stfD["NetID"]==netID)]["Fulfilled Substitutions"].values[0] return (APP,ACC,REJ,CAN,FUL) #Get a "high score" list or a "low score" list of names/scores based on a numeric string showing which categories to count #0: Approved, 1: Accepted, 2: Rejected, 3: Cancelled, 4: Fulfilled #E.g. "12" gives the ULAs with the most total accepted and rejected requests thusfar. def getTopSubs(categoryStr,fewest=False,num=1): if categoryStr=="": return "" stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) columns=[] if "0" in categoryStr: columns+=["Approved Substitutions"] if "1" in categoryStr: columns+=["Accepted Substitutions"] if "2" in categoryStr: columns+=["Rejected Substitutions"] if "3" in categoryStr: columns+=["Cancelled Substitutions"] if "4" in categoryStr: columns+=["Fulfilled Substitutions"] header="" for name in columns: header+=name[:3].upper()+"+" header=header[:-1] stfD[header]=0 for name in columns: stfD[header]+=(stfD[name].astype(float)) return stfD.sort_values(header,axis=0,ascending = fewest).head(num)[["Name",header]].to_string(index=False) #Test if a given single shift exists, i.e. if that section meets on that date. def shiftExists(period,date): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) try: temp=secD[(secD['Section']==period) & (secD['Date']==date)] temp.values[0] except IndexError: return False return True #Get a list of all names currently assigned to a given section on a given date def getAllNamesFromSection(periods,dates): secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) others=[] for period in periods.split(';'): for date in dates.split(';'): try: temp=secD[(secD['Section']==period) & (secD['Date']==date)] netIDs=list(temp.values[0][5:16]) netIDs = [IDToName(x) for x in netIDs if pd.notna(x)] others.append(cp.deepcopy(netIDs)) except IndexError: others.append("NoShift") return(others) #finds which date is earlier in string MM/DD format def dateBeforeOrEqual(date1, date2): month1,day1 = date1.split('/') month2,day2 = date2.split('/') month1=int(month1) month2=int(month2) day1=int(day1) day2=int(day2) if month1month2: return False elif day1<=day2: return True else: return False #Count how many subs each person *fulfilled* between two dates def getNumFulfilledInDateRange(subRequestsArchiveFilename,earlyDate,lateDate): kudos={} with open(subRequestsArchiveFilename) as f: headerLine=f.readline() reader = csv.reader(f) for row in reader: vals=row result = vals[6] if result not in ['APP','ACC']:#don't count if the request was rejected/cancelled, skip this line for performance reasons continue fulfiller=vals[4] if fulfiller == '':#Nobody fulfilled this request, skip this line for performance reasons continue #count how many 2-hour blocks were fulfilled dateCount=0 dates=vals[3].split(';') for date in dates: if dateBeforeOrEqual(earlyDate,date) and dateBeforeOrEqual(date,lateDate): dateCount+=1 #none of the dates are within the range, skip this line for performance reasons if dateCount<=0: continue #count how many 2-hour blocks were fulfilled sectionCount=0 sections=vals[2].split(';') for section in sections: if "HR" in section: sectionCount+=0.5 else: sectionCount+=1 if fulfiller in kudos: kudos[fulfiller]+=sectionCount*dateCount else: kudos[fulfiller]=sectionCount*dateCount return kudos #create string files for easy pasting into Google forms def createStrings(): stfD = pd.read_csv(staffDatabaseFilename,dtype=str,index_col=False) secD = pd.read_csv(sectionDatabaseFilename,dtype=str,index_col=False) with open("stringOutputs/staffIDs.txt",'w+') as f: netIDs=stfD['NetID'].values netIDs.sort() f.write('\n'.join(netIDs)) with open("stringOutputs/emails.txt",'w+') as f: vals=cp.deepcopy(stfD['NetID'].values) for i in range(len(vals)): vals[i]=vals[i]+"@msu.edu" f.write('; '.join(vals)) with open("stringOutputs/sectionTitles.txt",'w+') as f: vals=secD['Section'].values seen=set() f.write('\n'.join([x for x in vals if not (x in seen or seen.add(x))])) with open("stringOutputs/periodDates.txt",'w+') as f: vals=secD['Date'].values seen=set() f.write('\n'.join([x for x in vals if not (x in seen or seen.add(x))])) #Create the content of form-letter emails to be sent based on the request details and the chosen approval status def generateEmails(requestList,actuallySendEmails=False): emails=[] for req in requestList: #Unpack the request into named variables (for readability in the string formatting below) timestamp = req[0] requestor = IDToName(req[1]).strip("*") section = req[2] date = req[3] replacement = IDToName(req[4]).strip("*") reason = req[5] status = req[6] statusReason = req[7] #Check if a replacement was specified at all if replacement!="": replaced=True else: replaced=False #Unpack sections and dates to always be a list if ";" in section: sections = section.split(";") else: sections=[section] if ";" in date: dates = date.split(";") else: dates=[date] times=[] locations=[] if status=="APP" or status=="ACC": #For now, don't tell the ULAs that their acceptance was begrudging so they don't try and game the system by lying about the real reason. We can change this later if desired. status="approved" changed=True #Our database entries were changed for s in sections: times.append(getTimeFromSection(s)) locations.append(getLocationFromSection(s)) elif status=="REJ" or status=="CAN": if status == "REJ": status="rejected" else: status="cancelled" changed=False #Our database entries were changed for s in sections: try: times.append(getTimeFromSection(s)) except: times=["INVALID DATE/SECTION"] try: locations.append(getLocationFromSection(s)) except: locations=["INVALID DATE/SECTION"] else: print ('ERROR: Status of request is "'+str(status)+'" instead of APP/ACC/REJ/CAN') return recipient = req[1]+"@msu.edu" subject = "Your request for substitution on "+date+" has been "+status+"." sectionStrings=[] for i in range(len(sections)): if "HR" in sections[i]:#helproom sectionStrings.append(sections[i].replace("_HR"," helproom (hour ")+")") sectionStrings[i]=sectionStrings[i].replace("M","Monday") sectionStrings[i]=sectionStrings[i].replace("T","Tuesday") sectionStrings[i]=sectionStrings[i].replace("W","Wednesday") sectionStrings[i]=sectionStrings[i].replace("R","Thursday") sectionStrings[i]=sectionStrings[i].replace("F","Friday") sectionStrings[i]=sectionStrings[i].replace("S","Sunday") else: sectionStrings.append(sections[i].replace("Sec","Section")) message = "Hello "+requestor+"," message+="\n\nYour request to be substituted out of ["+" and ".join(sectionStrings)+"] at ["+" and ".join(times)+"] on ["+" and ".join(dates)+'] because of "'+reason+'" was recieved at ['+timestamp+"]." if replaced: message+="\nYou specified "+replacement+" as your replacement" if changed: message+=" and they have also recieved an automatically generated email notifying them of the switch." else: message+="." else: message+="\nYou did not specify a replacement." message+="\n\nThis request has been reviewed by the professors and *"+status+"*" if statusReason != "INST. REASON": message+=' for the given reason of: "'+statusReason+'"' if changed: message+=" and the corresponding changes have been made in our calander.\nIf all of the above information is correct, no further action is necessary. If any of this information is incorrect, please contact us IMMEDIATELY on Slack so we can correct it." else: message+= " and no changes have been made in our calander. If you did not submit this request, please contact us on Slack as there may be an error in our system." message+="\n\nThis email was automatically generated. Do not reply or send email to this address as it will not be seen. All communication should occur in Slack or by emailing cse102@msu.edu" emails.append([recipient,subject,message]) #Also send email to new ULA if their subbing IN was approved/accepted if changed and replaced: recipient = req[4]+"@msu.edu" subject = "You have been scheduled to substitute on "+date+"." message = "Hello "+replacement+"," message+="\n\nYou have been scheduled to substitute for "+requestor+" in ["+" and ".join(sectionStrings)+"] at ["+" and ".join(times)+"] on ["+" and ".join(dates)+"] at ["+" and ".join(locations)+"]" message+="\n\n"+requestor+" has specified that you have already agreed to this by submitting the request with you as a replacement and thus our calander has been changed." message+="\nIf all of the above information is correct, no further action is necessary. If any of this information is incorrect, please contact us IMMEDIATELY on Slack so we can correct it." message+="\n\nThis email was automatically generated. Do not reply or send email to this address as it will not be seen. All communication should occur in Slack or by emailing cse102@msu.edu" emails.append([recipient,subject,message]) #Also send email to GA if their helproom ULAs have been changed. if changed: for s in sections: if s in ["M_HR1","M_HR2","T_HR1","T_HR2","W_HR1","W_HR2","R_HR1","R_HR2","F_HR1","F_HR2","S_HR1","S_HR2"]: for date in dates: names=getAllNamesFromSection(s,date)[0] print(names,names!="NoShift") if names!="NoShift": recipient = nameToID(names[0])+"@msu.edu" subject="Change in helproom staff on "+date message="There has been a substitution in your helproom section: "+s+".\nYour ULAs on "+date+" are "+", ".join(names[1:]) emails.append([recipient,subject,message]) #Send emails (or print to terminal if debugging and actuallySend == False) sendEmails(emails,actuallySend=actuallySendEmails) #duumyParam is to make really sure that the author of the code that calls this functions really intends to actually send emails (by forcing them to use the keyword) def sendEmails(emails,dummyParam="DUMMY",actuallySend=False): #Catch incorrect use of this method by someone who doesn't understand the args. if dummyParam!="DUMMY": print("DON'T USE THE helperFunctions.sendEmails() method unless you know what you're doing! This can send actual emails to people's inbox!") return if actuallySend: #https://www.geeksforgeeks.org/send-mail-gmail-account-using-python/ # creates SMTP session s = smtplib.SMTP('smtp.gmail.com', 587) # start TLS for security s.starttls() # Authentication s.login("cse102msu@gmail.com", "whol ydag otqa hxps") print("BEGINNING EMAIL SENDING!") for email in emails: address = email[0] subject = email[1] message = email[2] messageObj = 'Subject: {}\n\n{}'.format(subject, message) # sending the mail s.sendmail("cse102msu@gmail.com", address, messageObj) # terminating the session print("Emails sent!") s.quit() else: print('\n!!!PRINTING TO TERMINAL INSTEAD TO PREVENT ACCIDENTALY EMAIL SPAM!!!:\n\t(Use the argument "acutallySend=True" in "sendEmails" function to disable this.)') for email in emails: print("\n"+"-"*80+"\nEmail would be sent to "+email[0]+":\nSubject: "+email[1]+"\nMessage:\n"+email[2]+"\n"+"-"*80) #Get the raw data from the Google Forms API #FYI: The first part of this code dealing with connection/authentication was not written by me and is largely a black box. #This function only adds requests that are more recent than the last time it was run, so that the original data in the Google cloud never needs to be cleared def getForms(subRequestsFilename): #BLACKBOX AUTHENTICATION MAGIC #-------------------------------------------------------------------------- #Requires some installation/setup https://developers.google.com/forms/api/quickstart/python SCOPES = ["https://www.googleapis.com/auth/forms.responses.readonly"] DISCOVERY_DOC = "https://forms.googleapis.com/$discovery/rest?version=v1" creds = None # The file token.json stores the user's access and refresh tokens, and is # created automatically when the authorization flow completes for the first # time. if os.path.exists('token.json'): creds = Credentials.from_authorized_user_file('token.json', SCOPES) store = file.Storage('token.json') if not creds or not creds.valid: if creds and creds.expired and creds.refresh_token: creds.refresh(Request()) else: flow = InstalledAppFlow.from_client_secrets_file('credentials.json', SCOPES) creds = flow.run_local_server(port=0) #Save the credentials for the next run with open('token.json', 'w+') as token: token.write(creds.to_json()) #service = discovery.build('forms', 'v1', http=creds.authorize(Http()), discoveryServiceUrl=DISCOVERY_DOC, static_discovery=False) service = build('forms', 'v1', credentials=creds) # gets the responses of your specified form: form_id = '1x-8fkuMAcQlTl36SdsbCG0tfClKAcvNshnV8L_Hl904' result = service.forms().responses().list(formId=form_id).execute() #END OF BLACKBOX AUTHENTICATION MAGIC #-------------------------------------------------------------------------- subs = pd.read_csv(subRequestsFilename,dtype=str) #Check when the last time the data was downloaded with open("lastUpdatedToken.txt",'r') as f: line = f.readline() if line != "": prevTime=datetime.strptime(line,"%Y-%m-%d %H:%M:%S.%f") else: prevTime=datetime.strptime("1975-01-01 01:01:01.000000","%Y-%m-%d %H:%M:%S.%f")#If the file was blank, (such as by just being created) use an aribtrary very early date data=result["responses"] #Unpack the ugly data structure that Google Forms returns for req in data: try: reason=req["answers"]["22a5ae9b"]["textAnswers"]["answers"][0]["value"] except KeyError:#No reason specified reason = "" requestor=req["answers"]["7bb6a9dd"]["textAnswers"]["answers"][0]["value"] dates="" for i in req["answers"]["11d3b4f8"]["textAnswers"]["answers"]: dates+=i["value"]+";" dates=dates[:-1] try: replacement=req["answers"]["30178530"]["textAnswers"]["answers"][0]["value"] except KeyError:#no replacement specified replacement="" sections="" for i in req["answers"]["5684403f"]["textAnswers"]["answers"]: sections+=i["value"]+";" sections=sections[:-1] timeStr = req["createTime"][:-1].replace("T"," ")+"000" #The timestamp needs formatting adjustment and a time-zone shift to EST timeStamp=datetime.strptime(timeStr, '%Y-%m-%d %H:%M:%S.%f')-timedelta(hours=4,minutes=0) #If the request is more recent than our last download, then our database doesn't yet 'know' about it and it needs to be added. if timeStamp>prevTime: reqDict={"Timestamp": [timeStr], "Requestor": [requestor], "Section": [sections], "Dates": [dates], "Replacement": [replacement], "Reason": [reason]} subs=pd.concat([subs,pd.DataFrame(reqDict)],ignore_index=True) subs.sort_values(by=["Timestamp"],inplace=True) #Write the updated request list to file subs.to_csv(subRequestsFilename,index=False,index_label=False) #Write the timestamp to the token for checking when this function was last run with open("lastUpdatedToken.txt",'w+') as f: f.write(str(datetime.now()))