repo
stringlengths
2
99
file
stringlengths
13
225
code
stringlengths
0
18.3M
file_length
int64
0
18.3M
avg_line_length
float64
0
1.36M
max_line_length
int64
0
4.26M
extension_type
stringclasses
1 value
Fanoos
Fanoos-master/experimentInfulstructure/randomlyGenerateQueriesToRun.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np; def ensures(booleanCondition): assert(booleanCondition); def requires(booleanCondition): assert(booleanCondition); def generateParameters(): dictToReturn = dict(); dictToReturn["numberOfDisjuncts"] = int(np.random.randint(1, 5)); dictToReturn["lengthConjuncts"] = [int(np.random.randint(1, 4)) for x in range(0, dictToReturn["numberOfDisjuncts"])]; dictToReturn["numberOfQuestionsPerQuestionType"] = 5; ensures(len(dictToReturn["lengthConjuncts"]) == dictToReturn["numberOfDisjuncts"]); return dictToReturn; def generateIndividualSentence(parameters, termList): requires(isinstance(parameters, dict)); requires("lengthConjuncts" in parameters); requires(isinstance(parameters["lengthConjuncts"], list)); requires([isinstance(x, int) for x in parameters["lengthConjuncts"]]); requires([ (x > 0) for x in parameters["lengthConjuncts"]]); stringToReturn = ""; maximumNumberOfIterations=1000; # to prevent infinite loop due to exhaustion of possiblities or # poor randomization. assert(maximumNumberOfIterations > 0); collectionOfTermsThusFar = set(); collectionOfTermsThusFar.add(frozenset()); # see the start of the while loop below. for thisConjunctLength in parameters["lengthConjuncts"]: if(len(stringToReturn) > 0): stringToReturn = stringToReturn + " or "; thisConjunct = frozenset(); tempIterationCount=maximumNumberOfIterations; while(thisConjunct in collectionOfTermsThusFar): thisConjunct = frozenset(np.random.choice(termList, thisConjunctLength, replace=False)); tempIterationCount = tempIterationCount -1; if(tempIterationCount <= 0): raise Exception("Exhausted maximum number of iterations. Sentence thus far:" + stringToReturn); collectionOfTermsThusFar.add(thisConjunct); assert(thisConjunct in collectionOfTermsThusFar); if(len(thisConjunct) > 1): stringToReturn = stringToReturn + "and( " + (" , ".join(thisConjunct)) + " )"; else: stringToReturn = stringToReturn + list(thisConjunct)[0]; return (stringToReturn, frozenset(collectionOfTermsThusFar)); import time; def generateSentences(parameters, dictMappingQuestionTypeToAvailablePredicates): requires(isinstance(parameters, dict)); requires("numberOfQuestionsPerQuestionType" in parameters); requires(isinstance(parameters["numberOfQuestionsPerQuestionType"], int)); requires(parameters["numberOfQuestionsPerQuestionType"] > 0); collectionOfSentencesFormed = set(); collectionOfSentencesFormed.add(frozenset()); maximumNumberOfIterations=1000; # to prevent infinite loop due to exhaustion of possiblities or # poor randomization. stringsToReturn = []; for thisQuestionType in dictMappingQuestionTypeToAvailablePredicates: for thisQuestionIndex in range(0, parameters["numberOfQuestionsPerQuestionType"]): nestedSetRepresentationForSentence = frozenset(); tempIterationCount = maximumNumberOfIterations; while(nestedSetRepresentationForSentence in collectionOfSentencesFormed): time.sleep(5 + np.random.randint(3) ); # To help increase the likelyhood that the randomizer will have a more # diverse selection, under the assumption that it in part depends on the time generation # occurs. np.random.seed(int( (time.clock() * 100000) % 10000) ); # In case the random generator uses the processor # time - which should not change during a call to time.sleep - as oppossed to the clock-time.... increaseDiversityInResponces = generateParameters(); stringForSentence, nestedSetRepresentationForSentence = generateIndividualSentence(\ increaseDiversityInResponces,\ dictMappingQuestionTypeToAvailablePredicates[thisQuestionType]); nestedSetRepresentationForSentence = frozenset([thisQuestionType, nestedSetRepresentationForSentence]); tempIterationCount = tempIterationCount -1; if(tempIterationCount <= 0): # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg raise Exception("Exhausted maximum number of iterations. Sentence thus far:" + stringToReturn); collectionOfSentencesFormed.add(nestedSetRepresentationForSentence); stringForSentence = thisQuestionType + " " + stringForSentence + " ?"; stringsToReturn.append(stringForSentence); ensures(len(stringsToReturn) == len(set(stringsToReturn))); # each sentence returned should be unique... return stringsToReturn; def getDictionariesToUse(): def convertDictEntries(thisDict): thisDict["what_do_you_ussually_do_when"] = thisDict["what_do_you_do_when"]; thisDict["when_do_you_ussually"] = thisDict["when_do_you"]; thisDict["what_are_the_usual_circumstances_in_which"] = thisDict["what_are_the_circumstances_in_which"]; return {x : [x for x in thisDict[x].replace("\n", "").split(" ") if len(x) > 0] for x in thisDict}; cpuDomainDict = {\ "when_do_you" : """ lwrite_high_ lwrite_very_low_ swrite_very_high_ usr_near_normal_levels writes_low lwrite_low_ swrite_high_ swrite_very_low_ usr_very_high_ writes_very_high lwrite_near_normal_levels swrite_low_ usr_high_ usr_very_low_ writes_very_low lwrite_very_high_ swrite_near_normal_levels usr_low_ writes_high""", \ "what_do_you_do_when" : """ free_space_high freemem_very_high_ lread_high_ reads_very_high sread_high_ free_space_low freemem_very_low_ lread_low_ reads_very_low sread_low_ free_space_very_high freeswap_high_ lread_near_normal_levels scall_high_ sread_near_normal_levels free_space_very_low freeswap_low_ lread_very_high_ scall_low_ sread_very_high_ freemem_high_ freeswap_near_normal_levels lread_very_low_ scall_near_normal_levels sread_very_low_ freemem_low_ freeswap_very_high_ reads_high scall_very_high_ freemem_near_normal_levels freeswap_very_low_ reads_low scall_very_low_ """, \ "what_are_the_circumstances_in_which" : """ free_space_high lread_low_ reads_very_high swrite_very_high_ free_space_low lread_near_normal_levels reads_very_low swrite_very_low_ free_space_very_high lread_very_high_ scall_high_ user_land_related_activity_high free_space_very_low lread_very_low_ scall_low_ user_land_related_activity_low freemem_high_ lwrite_high_ scall_near_normal_levels user_land_related_activity_very_high freemem_low_ lwrite_low_ scall_very_high_ user_land_related_activity_very_low freemem_near_normal_levels lwrite_near_normal_levels scall_very_low_ usr_high_ freemem_very_high_ lwrite_very_high_ sread_high_ usr_low_ freemem_very_low_ lwrite_very_low_ sread_low_ usr_near_normal_levels freeswap_high_ nonuser_system_activity_high sread_near_normal_levels usr_very_high_ freeswap_low_ nonuser_system_activity_low sread_very_high_ usr_very_low_ freeswap_near_normal_levels nonuser_system_activity_very_high sread_very_low_ writes_high freeswap_very_high_ nonuser_system_activity_very_low swrite_high_ writes_low freeswap_very_low_ reads_high swrite_low_ writes_very_high lread_high_ reads_low swrite_near_normal_levels writes_very_low""" \ }; cpuDomainDict = convertDictEntries(cpuDomainDict); invertedDoublePendulumDomain = { \ "when_do_you" : \ """outputtorque_high__magnitude outputtorque_magnitude_near_normal_levels statevalueestimate_low_ outputtorque_is_greater_than_or_equal_to_zero outputtorque_very_high__magnitude statevalueestimate_near_normal_levels outputtorque_is_less_than_or_equal_to_zero outputtorque_very_low__magnitude statevalueestimate_very_high_ outputtorque_low__magnitude statevalueestimate_high_ statevalueestimate_very_low_ """, \ "what_do_you_do_when" : \ """anchor_point_barely_moving_if_at_all pole1angle_rateofchange_high__magnitude pole2angle_rateofchange_very_high__magnitude anchor_point_moving_left pole1angle_rateofchange_low__magnitude pole2angle_rateofchange_very_low__magnitude anchor_point_moving_right pole1angle_rateofchange_magnitude_near_normal_levels pole2angle_very_high__magnitude both_poles_pointed_to_the_left pole1angle_rateofchange_very_high__magnitude pole2angle_very_low__magnitude both_poles_pointed_to_the_right pole1angle_rateofchange_very_low__magnitude pole_2_is_on_the_left_of_the_robot_chassy endofpole2_x_high__magnitude pole1angle_very_high__magnitude pole_2_is_on_the_right_of_the_robot_chassy endofpole2_x_low__magnitude pole1angle_very_low__magnitude poles_are_bent endofpole2_x_magnitude_near_normal_levels pole2_angle_at_of_above_x_axis_ poles_are_bent_like_the_arc_in_a_a_d endofpole2_x_very_high__magnitude pole2_close_to_vertical poles_are_bent_like_the_arc_in_a_c endofpole2_x_very_low__magnitude pole2_moving_barely poles_are_roughly_straight_in_respect_to_each_other pole1_angle_at_of_above_x_axis_ pole2_moving_clockwise_ vx_high__magnitude pole1_close_to_vertical pole2_moving_counter-clockwise_ vx_low__magnitude pole1_moving_barely pole2_on_left vx_magnitude_near_normal_levels pole1_moving_clockwise_ pole2_on_right vx_very_high__magnitude pole1_moving_counter-clockwise_ pole2angle_high__magnitude vx_very_low__magnitude pole1_on_left pole2angle_low__magnitude x_high__magnitude pole1_on_right pole2angle_magnitude_near_normal_levels x_low__magnitude pole1angle_high__magnitude pole2angle_rateofchange_high__magnitude x_magnitude_near_normal_levels pole1angle_low__magnitude pole2angle_rateofchange_low__magnitude x_very_high__magnitude pole1angle_magnitude_near_normal_levels pole2angle_rateofchange_magnitude_near_normal_levels x_very_low__magnitude """, \ "what_are_the_circumstances_in_which" : """ anchor_point_barely_moving_if_at_all pole1angle_magnitude_near_normal_levels pole_2_is_on_the_right_of_the_robot_chassy anchor_point_moving_left pole1angle_rateofchange_high__magnitude poles_are_bent anchor_point_moving_right pole1angle_rateofchange_low__magnitude poles_are_bent_like_the_arc_in_a_a_d both_poles_pointed_to_the_left pole1angle_rateofchange_magnitude_near_normal_levels poles_are_bent_like_the_arc_in_a_c both_poles_pointed_to_the_right pole1angle_rateofchange_very_high__magnitude poles_are_roughly_straight_in_respect_to_each_other endofpole2_x_high__magnitude pole1angle_rateofchange_very_low__magnitude speed_close_to_constant_assuming_no_friction endofpole2_x_low__magnitude pole1angle_very_high__magnitude speed_constant_assuming_no_friction endofpole2_x_magnitude_near_normal_levels pole1angle_very_low__magnitude speed_decreasing_assuming_no_friction endofpole2_x_very_high__magnitude pole2_angle_at_of_above_x_axis_ speed_increasing_assuming_no_friction endofpole2_x_very_low__magnitude pole2_close_to_vertical statevalueestimate_high_ outputtorque_high__magnitude pole2_moving_barely statevalueestimate_low_ outputtorque_is_greater_than_or_equal_to_zero pole2_moving_clockwise_ statevalueestimate_near_normal_levels outputtorque_is_less_than_or_equal_to_zero pole2_moving_counter-clockwise_ statevalueestimate_very_high_ outputtorque_low__magnitude pole2_on_left statevalueestimate_very_low_ outputtorque_magnitude_near_normal_levels pole2_on_right vx_high__magnitude outputtorque_very_high__magnitude pole2angle_high__magnitude vx_low__magnitude outputtorque_very_low__magnitude pole2angle_low__magnitude vx_magnitude_near_normal_levels pole1_angle_at_of_above_x_axis_ pole2angle_magnitude_near_normal_levels vx_very_high__magnitude pole1_close_to_vertical pole2angle_rateofchange_high__magnitude vx_very_low__magnitude pole1_moving_barely pole2angle_rateofchange_low__magnitude x_high__magnitude pole1_moving_clockwise_ pole2angle_rateofchange_magnitude_near_normal_levels x_low__magnitude pole1_moving_counter-clockwise_ pole2angle_rateofchange_very_high__magnitude x_magnitude_near_normal_levels pole1_on_left pole2angle_rateofchange_very_low__magnitude x_very_high__magnitude pole1_on_right pole2angle_very_high__magnitude x_very_low__magnitude pole1angle_high__magnitude pole2angle_very_low__magnitude pole1angle_low__magnitude pole_2_is_on_the_left_of_the_robot_chassy """\ } invertedDoublePendulumDomain = convertDictEntries(invertedDoublePendulumDomain); return {"invertedDoublePendulumDomain" : invertedDoublePendulumDomain, \ "cpuDomainDict" : cpuDomainDict }; dictsToUse = getDictionariesToUse(); commandSequencesToRunAfter= [\ "\n0.25\nq\nl\nq\nm\nq\nb\nexit\n", \ "\n0.125\nq\nm\nq\nl\nq\nb\nexit\n" \ ]; preamble = {\ "invertedDoublePendulumDomain" : "1\n./trainedNetworks/invertedDoublePendulumBulletEnv_v0/networkLayers_putIntoProperFormat.pickle\n0\n", \ "cpuDomainDict" : "2\n./trainedNetworks/cpuPolynomialRegressionModel/trainedPolynomialModelInfo.pickle\n1\n" \ }; listOfListsOfResults = []; for thisKey in dictsToUse: thisList = []; listOfListsOfResults.append(thisList); for thisString in generateSentences(generateParameters(), dictsToUse[thisKey]): for thisEndPart in commandSequencesToRunAfter: commandOverall = preamble[thisKey] + thisString + thisEndPart; thisList.append(commandOverall.replace("\n", "\\n")); assert(len(listOfListsOfResults[0]) == len(listOfListsOfResults[1])); # below is done to interleave the results from the two domains. coupledValues = list(zip(listOfListsOfResults[0], listOfListsOfResults[1])); randomPermutedIndices = np.random.permutation(len(listOfListsOfResults[0])); for thisIndex in randomPermutedIndices: print(coupledValues[thisIndex][0]); print(coupledValues[thisIndex][1]);
21,502
82.023166
2,828
py
Fanoos
Fanoos-master/utils/getPathToThisDirectory.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import os import inspect def getPathToThisDirectory(): """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ pathToThisDirectory = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))); return pathToThisDirectory + "/";
4,143
85.333333
2,781
py
Fanoos
Fanoos-master/utils/getGitCommitHash.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import os; from utils.getPathToThisDirectory import getPathToThisDirectory; import re; from utils.contracts import *; import uuid; def getGitCommitHash(UUID): requires(isinstance(UUID, str)); requires(re.match("^[0-9a-f\-]+$", UUID) != None); # we requires the UUID to # be a numeric string perhaps with underscores. We primarly require this # to ensure that a file name we generate below will be accepted by the # system - no space, no special character, etc....) pathToThisDirectory = getPathToThisDirectory(); assert(isinstance(pathToThisDirectory , str)); assert(len(pathToThisDirectory) > 0); assert(pathToThisDirectory[0] == "/"); # we require the path name an absolute path assert(pathToThisDirectory[-1] == "/"); # we require that pathToThisDirectory specify a directory. # Below we name the file with the UUID so that if multiple processes are running this same # code in the same directory, they do not conflict over generating then deleting the file. pathToFileToSaveCommitInfo = pathToThisDirectory + "tempForGitHash" + UUID +".txt" # tempted to make the file not visible with .tempForGitHash successOfQuery = os.system("git log -n1 > " + pathToFileToSaveCommitInfo); if(successOfQuery != 0): # zero means no error return None; assert(successOfQuery == 0); fh = open(pathToFileToSaveCommitInfo, "r"); firstLine = fh.readline(); fh.close(); assert(isinstance(firstLine, str)); assert(len(firstLine) > 8); # at least 7 characters for the beginning "commit " # and one character for the ending newline. assert(firstLine[0:7] == "commit "); assert(firstLine[-1] == "\n"); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ proposedCommitHash = firstLine[7:-1]; # below assert is checking that we extracted the full hash in the range assert("commit " + proposedCommitHash + "\n" == firstLine); # the below assert checks that the proposed hash is an alphanumeric string from # beginning to end assert(re.match("^[a-zA-Z0-9]*$", proposedCommitHash) != None); os.system("rm " + pathToFileToSaveCommitInfo); # removing the temporary file after having # read it. return proposedCommitHash; gitCommitHashWhenThisCodeStartedRunning = getGitCommitHash(str(uuid.uuid4()));
6,249
65.489362
2,781
py
Fanoos
Fanoos-master/utils/contracts.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); """ A brief note on contracts and their proper/intended use: If you want to check that the output of some ESSENTIAL code meets some condition, then the way to do that is: (a) resultName = essentialCode ; assert( check(resultName) ); OR (b) resultName = essentialCode ; if( check(resultName) ): raise Exception(errorMessage); the difference between (a) and (b) is that (a) performs a check and produces an error that IS NOT part of the intended behavior of the code, while (b) produces an error that IS an intended behavior of the code. Essentially, THE INTENDED BEHAVIOR OF THE CODE SHOULD NOT RELY ON CONTRACTS - IT SHOULD BE POSSIBLE TO REMOVE ALL CONTRACTS AND FOR THE CODE TO FUNCTION EXACTLY THE SAME WAY IN ALL RELEVANT CIRCUMSTANCES, INCLUDING IN ANY PLANNED HANDLING OF USER-INPUT ERRORS, ETC. Thus, if one wants to produce robust output for incorrect user inputs when deployed, (b) is the correct way to go. However, if one is testing, concerned with checking internal correctness, and/or is not focused on covering the deployment cases of user-input errors, (a) might not be inappropriate. Requires, ensures, and asserts, while they may evaluate to the same results, do not convey the same information. In brief, requires are meant to convey function preconditions, asserts are meant to express expected behavior at a specific location assuming preconditions hold, and ensures are intended to convey the guaranteed postconditions provided that the requires are met. Some people sub-divide and expand contracts further, particularly in regard to statements about program invariants versus function arguments and return values --- I spare comments on that here, though I note the temptation to expand the number of contracts implemented in this file. For those interested by this concept, and especially for those who think it is silly, I suggest investigating the literature available on this sort of tool utilization / design philosophy / design approach ("contract programming", "design by contract", etc.). This comment is by no means intended to be a comprehensive treatment of contracts and their roles. If you want to remove the contracts from the code to squeeze out a bit more efficiency, please see the script <root of git repo>/removePythonContracts.py kept in the development branch. We highly recommend against removing contracts for any sort of casual use. Also, note that Python's runtime -O flag most likely will fail to optimize-out calls to requires and ensures as implemented below; a call of form "requires(<expensive call evaluating to bool>)" will still calculate the content in the parenthesis, even though "assert(<expensive call evaluating to bool>)" will not ( necessitating the script we provide). """ def requires(booleanStatement): assert(booleanStatement); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ def ensures(booleanStatement): assert(booleanStatement);
6,865
66.313725
2,781
py
Fanoos
Fanoos-master/utils/getStringTimeNow.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from datetime import datetime; def getStringTimeNow(): dateTimeString = str(datetime.now()); timeFieldSeperator = "_" specialCharactersToRemove = ['-',' ', '.', ":"]; for thisSpecialCharacter in specialCharactersToRemove: dateTimeString = dateTimeString.replace(thisSpecialCharacter, timeFieldSeperator); assert(all([ dateTimeString.count(x) == 0 for x in specialCharactersToRemove])); """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ return dateTimeString;
4,390
85.098039
2,781
py
Fanoos
Fanoos-master/utils/distributionStatics.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); from utils.contracts import *; import numpy as np; def distributionStatics(thisData): # unfortunately, some of the way this is used in the descriptionState (which I have # to retro-fit this into) makes it hard to write requires that won't break. # at the very least, thisData should be a single numerical value or some sort of # iterable which allows access to numerical items.... if(len(thisData) <= 0): return {"numberOfDataPoints" : 0}; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ else: return {\ "numberOfDataPoints" : len(thisData), "sumOfValues" : np.sum(thisData, axis=0), "median" : np.median(thisData, axis=0), "mean" : np.mean(thisData, axis=0), "std" : np.std(thisData, axis=0), "min" : np.min(thisData, axis=0), "max" : np.max(thisData, axis=0), "0.05quantile" : np.quantile(thisData, 0.05, axis=0), "0.10quantile" : np.quantile(thisData, 0.10, axis=0), "0.25quantile" : np.quantile(thisData, 0.25, axis=0), "0.75quantile" : np.quantile(thisData, 0.75, axis=0), "0.90quantile" : np.quantile(thisData, 0.90, axis=0), "0.95quantile" : np.quantile(thisData, 0.95, axis=0), }; raise Exception("Control should not reach here"); return;
5,309
73.788732
2,789
py
Fanoos
Fanoos-master/utils/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
Fanoos
Fanoos-master/CEGARLikeAnalysis/labelsForBoxes.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); # Below, we make the labels powers of two so we may use them as bit-flags. Recall # that because we have allocated only one byte for box-labels, this allows for only # eight flags. TRUEEVERYWHERE = 1; LOWESTLEVEL_FALSESOMEWHEREANDEXHAUSTEDLOOKING=2; # by lowest level, we mean smallest size, which actually # corresponds to largest depth in the CEGARLikeAnalysis ... # ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg HIGHESTLEVELBOXUNIONBOX_FALSESOMEWHEREANDEXHAUSTEDLOOKING = 4; FALSEEVERYWHERE = 8;
4,392
92.468085
2,783
py
Fanoos
Fanoos-master/CEGARLikeAnalysis/CEGARLikeAnalysisMain.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__); import numpy as np from utils.contracts import *; from propagateBoxThroughLearnedSystem.classesToPropogateBoxThroughModels import ModelBoxProgatorManager; from boxesAndBoxOperations.getBox import isProperBox, getBox, getDimensionOfBox, getJointBox, getContainingBox, getRandomBox; from boxesAndBoxOperations.CEGARFileWrittingManager import CEGARFileWrittingManager; def getInitialAbstraction_boxesBySign(universeBox): requires(isProperBox(universeBox)); requires(np.all(universeBox[:, 0] < universeBox[:, 1])); numberOfBoxes = 2 ** (universeBox.shape[0]); listToReturn = []; for thisboxIndex in range(0, numberOfBoxes): thisBox = universeBox.copy(); binaryRepresentationOfIndex = np.binary_repr(thisboxIndex, width=getDimensionOfBox(universeBox)); assert(isinstance(binaryRepresentationOfIndex, str)); assert(all([(x in {"1", "0"}) for x in binaryRepresentationOfIndex])); for thisVariableIndex in range(0, getDimensionOfBox(universeBox)): thisBox[thisVariableIndex, int(binaryRepresentationOfIndex[thisVariableIndex])] = np.mean(thisBox[thisVariableIndex, :]); listToReturn.append(thisBox); return listToReturn; from boxesAndBoxOperations.splitBox import splitBox; import inspect; import time; timingInfoForLocation_2e048534_BoxTest = []; from CEGARLikeAnalysis import labelsForBoxes; def helper_formDummyBoxes(thisInputBox, functionToDetermineWhenToGiveUpOnBox, CEGARFileWrittingManagerInstance, scalingForSplitting, depth): if(functionToDetermineWhenToGiveUpOnBox(thisInputBox)): CEGARFileWrittingManagerInstance.writeBox(thisInputBox, \ [depth, (labelsForBoxes.FALSEEVERYWHERE | labelsForBoxes.LOWESTLEVEL_FALSESOMEWHEREANDEXHAUSTEDLOOKING) ]); return; dummyBoxesToWrite = splitBox(thisInputBox, "randomNumberOfUniformSplits", scalingFactors=scalingForSplitting); for thisNextBox in dummyBoxesToWrite: helper_formDummyBoxes(thisNextBox, functionToDetermineWhenToGiveUpOnBox, CEGARFileWrittingManagerInstance, scalingForSplitting, depth + 1) return; def analysis_divingIntoBox(thisInputBox, thisInstanceOfModelBoxProgatorManager, functionToStatisfy, functionToDetermineWhenToGiveUpOnBox, \ CEGARFileWrittingManagerInstance, scalingForSplitting, depth, functionToCheckWhetherNoPointsInTheBoxStatisfyCondition=None): thisOutputBox = thisInstanceOfModelBoxProgatorManager.pushBoxThrough(thisInputBox); startTime = time.process_time(); #location2e048534-c79b-4177-a79d-cc0ef71384d4_boxTest boxTest = functionToStatisfy(thisInputBox, thisOutputBox); endTime = time.process_time(); timingInfoForLocation_2e048534_BoxTest.append(endTime - startTime); if(boxTest): CEGARFileWrittingManagerInstance.writeBox(thisInputBox, [depth, labelsForBoxes.TRUEEVERYWHERE]); return True; # True for success... elif(functionToDetermineWhenToGiveUpOnBox(thisInputBox)): CEGARFileWrittingManagerInstance.writeBox(thisInputBox, \ [depth, labelsForBoxes.LOWESTLEVEL_FALSESOMEWHEREANDEXHAUSTEDLOOKING]); return False; #False for failure. elif( (functionToCheckWhetherNoPointsInTheBoxStatisfyCondition is not None) and \ (functionToCheckWhetherNoPointsInTheBoxStatisfyCondition(thisInputBox, thisOutputBox)) ): CEGARFileWrittingManagerInstance.writeBox(thisInputBox, \ [depth, labelsForBoxes.FALSEEVERYWHERE]); helper_formDummyBoxes(thisInputBox, functionToDetermineWhenToGiveUpOnBox, \ CEGARFileWrittingManagerInstance, scalingForSplitting, depth); # yes, we pass the depth here, not depth+1, # because the further splitting occurs in helper_formDummyBoxes. return False; #False for failure. # goes up a layer to accumulate a larger box where all members of the box failed else: refimentElements = splitBox(thisInputBox, "randomNumberOfUniformSplits", scalingFactors=scalingForSplitting); resultOfFurtherRefining = [\ analysis_divingIntoBox(x, thisInstanceOfModelBoxProgatorManager, functionToStatisfy, functionToDetermineWhenToGiveUpOnBox, \ CEGARFileWrittingManagerInstance, scalingForSplitting, depth+1)\ for x in refimentElements]; if(not any(resultOfFurtherRefining)): return False; # goes up a layer to accumulate a larger box where all members of the box failed else: for thisIndex in range(0, len(resultOfFurtherRefining)): if(resultOfFurtherRefining[thisIndex]): # box succeeded and thus was written out by lower level... continue; CEGARFileWrittingManagerInstance.writeBox(refimentElements[thisIndex], \ [depth+1, labelsForBoxes.HIGHESTLEVELBOXUNIONBOX_FALSESOMEWHEREANDEXHAUSTEDLOOKING]); return True; # i.e., not all members of the box failed, so thisInputBox should not be accumulated into # a box where we label all members of the box as failures... raise Exception("Control should never reach here"); return; """ ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAAIAQC/RPJH+HUB5ZcSOv61j5AKWsnP6pwitgIsRHKQ5PxlrinTbKATjUDSLFLIs/cZxRb6Op+aRbssiZxfAHauAfpqoDOne5CP7WGcZIF5o5o+zYsJ1NzDUWoPQmil1ZnDCVhjlEB8ufxHaa/AFuFK0F12FlJOkgVT+abIKZ19eHi4C+Dck796/ON8DO8B20RPaUfetkCtNPHeb5ODU5E5vvbVaCyquaWI3u/uakYIx/OZ5aHTRoiRH6I+eAXxF1molVZLr2aCKGVrfoYPm3K1CzdcYAQKQCqMp7nLkasGJCTg1QFikC76G2uJ9QLJn4TPu3BNgCGwHj3/JkpKMgUpvS6IjNOSADYd5VXtdOS2xH2bfpiuWnkBwLi9PLWNyQR2mUtuveM2yHbuP13HsDM+a2w2uQwbZgHC2QVUE6QuSQITwY8RkReMKBJwg6ob2heIX+2JQUniF8GKRD7rYiSm7dJrYhQUBSt4T7zN4M5EDg5N5wAiT5hLumVqpAkU4JeJo5JopIohEBW/SknViyiXPqBfrsARC9onKSLp5hJMG1FAACezPAX8ByTOXh4r7rO0UPbZ1mqX1P6hMEkqb/Ut9iEr7fR/hX7WD1fpcOBbwksBidjs2rzwurVERQ0EQfjfw1di1uPR/yzLVfZ+FR2WfL+0FJX/sCrfhPU00y5Q4Te8XqrJwqkbVMZ8fuSBk+wQA5DZRNJJh9pmdoDBi/hNfvcgp9m1D7Z7bUbp2P5cQTgay+Af0P7I5+myCscLXefKSxXJHqRgvEDv/zWiNgqT9zdR3GoYVHR/cZ5XpZhyMpUIsFfDoWfAmHVxZNXF0lKzCEH4QXcfZJgfiPkyoubs9UDI7cC/v9ToCg+2SkvxBERAqlU4UkuOEkenRnP8UFejAuV535eE3RQbddnj9LmLT+Y/yRUuaB2pHmcQ2niT1eu6seXHDI1vyTioPCGSBxuJOciCcJBKDpKBOEdMb1nDGH1j+XpUGPtdEWd2IisgWsWPt3OPnnbEE+ZCRwcC3rPdyQWCpvndXCCX4+5dEfquFTMeU9LOnOiB1uZbnUez4AuicESbzR522iZZ+JdBk3bWyah2X8LW2QKP0YfZNAyOIufW4xSUCBljyIr9Z1/KhBFSMP2yibWDnOwQcK91Vh76AqmvaviTbZn9BrhzgndaODtWAyXtrWZX2iwo3lMpcx8qh3V9YeRB7sOYQVbtGhgDlY2jYv8fPWWaYGrNVvRm+vWUiSKdBgLR5mF0B/r7gC3FERNVecEHE1sMHIZmbd77QnGP9qlv/pP9x1RMHZVsvpSuAufaf6vqXQa5VwKEAt6CQwy7SpfTpBIcvH2qbSfVqPVewZ7ISg7UU+BvKZR5bwzTZSaLC2P4oPPAXeLCDDlC7+OFk3bJ/4Bq6v3NoqYh5d6o4C2lARUTYrwspWHrOTnd/4Osf3/YStqJ+CqdOxmu0xiX8bH+EJek5prI86iGYAJHttMFZcfXK+AJ2SOAJ0YIiV0YgQaeVc75KkNsRE6+mYjE1HZXKi6+wyHLSoJTGUv1WEpUdbGYJO32LVCGwDtG1qcSyVOgieHEwqB5W1qlZeoKLPUHWmziD09ojEsZurRtUKrvSGX/pwrKpDX2U229hJWXrTp13ZNHDdsLz+Brb8ZyGUb/o1aydw7O3ERvmB8drOeUP6PGgCkI26VjKIIEqXfTf8ciG1mssVcQolxNQT/ZZjo4JbhBpX+x6umLz3VDlOJNDnCXAK/+mmstw901weMrcK1cZwxM8GY2VGUErV3dG16h7CqRJpTLn0GxDkxaEiMItcPauV0g10VWNziTaP/wU3SOY5jV0z2WbmcZCLP40IaXXPL67qE3q1x/a18geSFKIM8vIHG8xNlllfJ60THP9X/Kj8GDpQIBvsaSiGh8z3XpxyuwbQIt/tND+i2FndrM0pBSqP8U3n7EzJfbYwEzqU9fJazWFoT4Lpv/mENaFGFe3pgUBv/qIoGqv2/G5u0RqdtToUA6gR9bIdiQpK3ZSNRMM2WG/rYs1c6FDP8ZGKBh+vzfA1zVEOKmJsunG0RU9yinFhotMlix14KhZMM6URZpDGN+zZ9lWMs6UMbfAwHMM+2MqTo6Se7var7uY5GDNXxQ9TTfDAWQw7ZAyzb0UR8kzQmeKrFbcPQ7uaIqV+HC4hj8COCqb/50xy6ZMwKVccw0mhVSt1NXZgoa6mx6cx251G9crWvxfPpvuYLH2NqnceoeADP8hTiia6N6iN3e4kBzDXHIrsgI6NFd6qW9p9HrFnDmHdakv3qfCJSY8acYdEe9ukRXvheyKGtvqmbMnS2RNDLcMwSQo9aypSPNpHMEXtvVp+vIuiWCR1fjgz8uY1f1Pa0SETX9jrLXfqq1zGeQTmFPR1/ANUbEz25nFIkwSUTr5YduvbFIruZ5cW8CySfKyiun+KclIwKhZVbHXcALjAOc//45HV0gdJfEEnhbUkQ+asWdf3Guyo6Eqd8g40X6XsJiFY5ah7Mc4IacNBzp3cHU3f0ODVjP9xTMMH+cNxq9IYvvhlVp38e8GydYCGoQ79jvKWHLbtsF+Z1j98o7xAxdBRKnCblSOE4anny07LCgm3U18Qft0HFEpIFATnLb3Yfjsjw1sE8Rdj9FBFApVvA3SvjGafvq5b7J9QnTWy80TjwL5zrix6vwxxClT/zjDNX+3PPXVr1FMF+Rhel58tJ8pMQ3TrzC1961GAp5eiYA1zGSyDPz+w== abc@defg """ from databaseInterface.databaseValueTracker import ObjDatabaseValueTracker; import pickle; def analysis(universeBox, thisInstanceOfModelBoxProgatorManager, functionToStatisfy, functionToDetermineWhenToGiveUpOnBox, \ limitSplittingToAxisWithIndicesInThisList=None, functionToCheckWhetherNoPointsInTheBoxStatisfyCondition=None): timingInfoForLocation_2e048534_BoxTest = []; requires(isinstance(limitSplittingToAxisWithIndicesInThisList, list)); requires( \ np.all([(x >= 0 and x < getDimensionOfBox(universeBox)) for x in limitSplittingToAxisWithIndicesInThisList])); requires( \ (len(set(limitSplittingToAxisWithIndicesInThisList)) == len(limitSplittingToAxisWithIndicesInThisList)) ); CEGARFileWrittingManagerInstance = CEGARFileWrittingManager(universeBox); CEGARFileWrittingManagerInstance.writeMetadata(\ "fileNameOfLoadedModel", thisInstanceOfModelBoxProgatorManager.fileNameOfLoadedModel); CEGARFileWrittingManagerInstance.writeMetadata(\ "functionToStatisfy", inspect.getsource(functionToStatisfy)); CEGARFileWrittingManagerInstance.writeMetadata(\ "functionToDetermineWhenToGiveUpOnBox", inspect.getsource(functionToDetermineWhenToGiveUpOnBox)); CEGARFileWrittingManagerInstance.writeMetadata(\ "QAStateUUID_mostRecentBeingComputed", ObjDatabaseValueTracker.get_QAStateUUID_mostRecentBeingComputed()); theseInputAbstractions = getInitialAbstraction_boxesBySign(universeBox); assert(isinstance(theseInputAbstractions, list)); assert(len(theseInputAbstractions) > 0); scalingForSplitting = universeBox[:, 1] - universeBox[:, 0]; tempBox = scalingForSplitting.copy(); # TODO: remove this unnecessary copy in the near future. # As implemented in the splitBox file, when the scaling factor has a nan in a posotion, # the axis corresponding to that index is ignored. tempBox[:] = np.nan; tempBox[limitSplittingToAxisWithIndicesInThisList] = scalingForSplitting[limitSplittingToAxisWithIndicesInThisList]; scalingForSplitting = tempBox; for thisBox in theseInputAbstractions: anySuccess = analysis_divingIntoBox(thisBox, thisInstanceOfModelBoxProgatorManager, \ functionToStatisfy, functionToDetermineWhenToGiveUpOnBox, CEGARFileWrittingManagerInstance, scalingForSplitting, 0, \ functionToCheckWhetherNoPointsInTheBoxStatisfyCondition=functionToCheckWhetherNoPointsInTheBoxStatisfyCondition); if(not anySuccess): CEGARFileWrittingManagerInstance.writeBox(thisBox, [0, labelsForBoxes.HIGHESTLEVELBOXUNIONBOX_FALSESOMEWHEREANDEXHAUSTEDLOOKING]); CEGARFileWrittingManagerInstance.closeFilesToSaveResultsIn(); return CEGARFileWrittingManagerInstance;
11,844
63.375
2,781
py
Fanoos
Fanoos-master/CEGARLikeAnalysis/__init__.py
# Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems ; David Bayani and Stefan Mitsch ; paper at https://arxiv.org/abs/2006.12453 # Copyright (C) 2021 David Bayani # # This file is part of Fanoos. # # Fanoos is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, version 3 of the License only. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # # Contact Information: # # Electronic Mail: # dcbayani@alumni.cmu.edu # # Paper Mail: # David Bayani # Computer Science Department # Carnegie Mellon University # 5000 Forbes Ave. # Pittsburgh, PA 15213 # USA # # import config; _LOCALDEBUGFLAG = config.debugFlags.get_v_print_ForThisFile(__file__);
1,154
28.615385
165
py
IJCAI2016
IJCAI2016-master/ConfusionMatrix.py
import numpy as np class ConfusionMatrix: """ Simple confusion matrix class row is the true class, column is the predicted class """ def __init__(self, n_classes): self.n_classes = n_classes self.mat = np.zeros((n_classes,n_classes),dtype='int') def __str__(self): return np.array_str(self.mat) def batchAdd(self,y_true,y_pred): assert len(y_true) == len(y_pred) assert max(y_true) < self.n_classes assert max(y_pred) < self.n_classes for i in range(len(y_true)): self.mat[y_true[i],y_pred[i]] += 1 def zero(self): self.mat.fill(0) def getErrors(self): """ Calculate differetn error types :return: vetors of true postives (tp) false negatives (fn), false positives (fp) and true negatives (tn) pos 0 is first class, pos 1 is second class etc. """ tp = np.asarray(np.diag(self.mat).flatten(),dtype='float') fn = np.asarray(np.sum(self.mat, axis=1).flatten(),dtype='float') - tp fp = np.asarray(np.sum(self.mat, axis=0).flatten(),dtype='float') - tp tn = np.asarray(np.sum(self.mat)*np.ones(self.n_classes).flatten(),dtype='float') - tp - fn - fp return tp,tn,fp,fn def accuracy(self): """ Calculates global accuracy :return: accuracyn :example: >>> conf = ConfusionMatrix(3) >>> conf.batchAdd([0,0,1],[0,0,2]) >>> print conf.accuracy() """ tp, _, _, _ = self.getErrors() n_samples = np.sum(self.mat) return np.sum(tp) / n_samples def sensitivity(self): tp, tn, fp, fn = self.getErrors() res = tp / (tp + fn) res = res[~np.isnan(res)] return res def specificity(self): tp, tn, fp, fn = self.getErrors() res = tn / (tn + fp) res = res[~np.isnan(res)] return res def positivePredictiveValue(self): tp, tn, fp, fn = self.getErrors() res = tp / (tp + fp) res = res[~np.isnan(res)] return res def negativePredictiveValue(self): tp, tn, fp, fn = self.getErrors() res = tn / (tn + fn) res = res[~np.isnan(res)] return res def falsePositiveRate(self): tp, tn, fp, fn = self.getErrors() res = fp / (fp + tn) res = res[~np.isnan(res)] return res def falseDiscoveryRate(self): tp, tn, fp, fn = self.getErrors() res = fp / (tp + fp) res = res[~np.isnan(res)] return res def F1(self): tp, tn, fp, fn = self.getErrors() res = (2*tp) / (2*tp + fp + fn) res = res[~np.isnan(res)] return res def matthewsCorrelation(self): tp, tn, fp, fn = self.getErrors() numerator = tp*tn - fp*fn denominator = np.sqrt((tp + fp)*(tp + fn)*(tn + fp)*(tn + fn)) res = numerator / denominator res = res[~np.isnan(res)] return res def getMat(self): return self.mat
3,077
28.596154
112
py
noelle
noelle-master/external/autotuner/src/autotuner.py
#!/usr/bin/env python import os import sys import json import traceback import opentuner from opentuner import ConfigurationManipulator from opentuner import IntegerParameter from opentuner.search.manipulator import BooleanParameter from opentuner import LogIntegerParameter from opentuner import SwitchParameter from opentuner import MeasurementInterface from opentuner import Result thisPath = os.path.dirname(os.path.abspath(__file__)) sys.path.append(thisPath + "/../utils") import utils techniqueIndexConverter = [utils.Technique.DOALL, utils.Technique.HELIX, utils.Technique.DSWP] class autotuneProgram(MeasurementInterface): ranges = None loopIDs = None confFile = None executionTimeFile = None exploredConfs = {} def getArgs(self): # Read the range of each dimension of the design space spaceFile = os.environ['autotunerSPACE_FILE'] self.ranges, self.loopIDs = utils.readSpaceFile(spaceFile) # Get autotuner.info self.confFile = os.environ['INDEX_FILE'] # Get execution time file self.executionTimeFile = os.environ['autotunerEXECUTION_TIME'] self.baselineTimeFile = os.environ['autotunerBASELINE_TIME'] self.baselineTime = utils.readExecutionTimeFile(self.baselineTimeFile) return def manipulator(self): """ Define the search space by creating a ConfigurationManipulator """ sys.stderr.write('AUTOTUNER: create the design space\n') self.getArgs() # Describe the design space to opentuner param = 0 manipulator = ConfigurationManipulator() for loopID in self.loopIDs: for elem in self.ranges[loopID]: dimensionIsDigit = elem.isdigit() dimension = 0 if (dimensionIsDigit): dimension = int(elem) # Check if the current dimension has a cardinality higher than 1 if ((dimension > 1) or (not dimensionIsDigit)): # Check the type of the parameter paramType = param % 9 # Create the parameter if (paramType == 0): start = 0 stop = dimension - 1 # Check if we want to force a specific parallelization technique forceDimension = len(elem.split("_")) == 2 if (forceDimension): _, considerLoop = elem.split("_") start = int(considerLoop) stop = int(considerLoop) # Should the loop be parallelized? openTuner_param = IntegerParameter(str(param), start, stop) elif (paramType == 1): # Unroll factor openTuner_param = IntegerParameter(str(param), 0, dimension - 1) elif (paramType == 2): # Peel factor openTuner_param = IntegerParameter(str(param), 0, dimension - 1) elif (paramType == 3): start = 0 stop = dimension - 1 # Check if we want to force a specific parallelization technique forceDimension = len(elem.split("_")) == 2 if (forceDimension): _, technique = elem.split("_") start = int(technique) stop = int(technique) # Parallelization technique openTuner_param = IntegerParameter(str(param), start, stop) # ED: should this be a SwitchParameter? elif (paramType == 4): # Number of cores to dedicate to the current loop openTuner_param = IntegerParameter(str(param), 0, dimension - 1) elif (paramType == 5): # DOALL parameter: chunk factor openTuner_param = IntegerParameter(str(param), 0, dimension - 1) elif (paramType == 6): # HELIX parameter: should we fix the maximum number of sequential segments? openTuner_param = SwitchParameter(str(param), dimension) elif (paramType == 7): # HELIX parameter: maximum number of sequential segments openTuner_param = IntegerParameter(str(param), 0, dimension - 1) elif (paramType == 8): # DSWP parameter: should we use queue packing? openTuner_param = SwitchParameter(str(param), dimension) # Share the parameter to OpenTuner manipulator.add_parameter(openTuner_param) param += 1 return manipulator def getNormalizedConf(self, confArg): conf = confArg.copy() startLoopIndex = 0 isLoopEnabled = False keysAsIntSorted = sorted([int(key) for key in confArg.keys()]) for key in keysAsIntSorted: value = conf[str(key)] if ((key % 9) == 0): startLoopIndex = key if (value == 0): # Loop is disabled isLoopEnabled = False for keyToSetToZero in range(key, key + 9): if (str(keyToSetToZero) in conf): conf[str(keyToSetToZero)] = 0 else: isLoopEnabled = True if (not isLoopEnabled): continue if ((key - startLoopIndex) == 3): conf[str(key)] = techniqueIndexConverter[value] if (conf[str(key)] != utils.Technique.DOALL): # DOALL was not chosen chunkSizeIndex = key + 2 conf[str(chunkSizeIndex)] = 0 return conf def getExpandedConf(self, confArg): conf = confArg.copy() index = 0 for loopID in self.loopIDs: for elem in range(0, len(self.ranges[loopID])): if (str(index) not in conf): conf[str(index)] = 0 index += 1 return conf def getConfWithLoopIDs(self, confArg): conf = {} for loopID in self.loopIDs: conf[loopID] = [] loopIDIndex = -1 keysAsIntSorted = sorted([int(key) for key in confArg.keys()]) for key in keysAsIntSorted: if ((key % 9) == 0): loopIDIndex += 1 loopID = self.loopIDs[loopIDIndex] conf[loopID].append(confArg[str(key)]) return conf def getConfAsStr(self, confArg): confAsStr = '' keysAsIntSorted = sorted([int(key) for key in confArg.keys()]) for key in keysAsIntSorted: confAsStr += str(confArg[str(key)]) + '_' return confAsStr def run(self, desired_result, input, limit): """ Compile and run a given configuration then return performance """ # Read the configuration to run conf = desired_result.configuration.data sys.stderr.write("AUTOTUNER: conf " + str(conf) + "\n") confNormalized = self.getNormalizedConf(conf) sys.stderr.write("AUTOTUNER: confNormalized " + str(confNormalized) + "\n") confExpanded = self.getExpandedConf(confNormalized) sys.stderr.write("AUTOTUNER: confExpanded " + str(confExpanded) + "\n") confExpandedAsStr = self.getConfAsStr(confExpanded) sys.stderr.write("AUTOTUNER: confExpandedAsStr " + str(confExpandedAsStr) + "\n") time = None # Check if configuration has already been run if (confExpandedAsStr in self.exploredConfs): time = self.exploredConfs[confExpandedAsStr] return Result(time = time) try: # Compile confWithLoopIDs = self.getConfWithLoopIDs(confExpanded) sys.stderr.write("AUTOTUNER: confWithLoopIDs " + str(confWithLoopIDs) + "\n") compileRetCode = utils.myCompile(self.confFile, confWithLoopIDs) if (compileRetCode != 0): time = float('inf') return Result(time = time) # Run parallel optimized binary maxExecutionTime = 2*self.baselineTime runRetCode = utils.myRun(maxExecutionTime) if (runRetCode != 0): time = float('inf') return Result(time = time) except KeyboardInterrupt: sys.stderr.write("AUTOTUNER: KeyboardInterrupt. Abort.\n") sys.exit(1) # Get execution time time = utils.readExecutionTimeFile(self.executionTimeFile) # Save conf in our list of explored configurations self.exploredConfs[confExpandedAsStr] = time return Result(time=time) def writeJson(self, pathToFile, jsonData): with open(pathToFile, 'w') as f: json.dump(jsonData, f) f.close() return def save_final_config(self, configuration): # Generate bitcode and binary with final configuration conf = configuration.data confNormalized = self.getNormalizedConf(conf) confExpanded = self.getExpandedConf(confNormalized) confExpandedAsStr = self.getConfAsStr(confExpanded) confWithLoopIDs = self.getConfWithLoopIDs(confExpanded) compileRetCode = utils.myCompile(self.confFile, confWithLoopIDs) if (compileRetCode != 0): sys.stderr.write("AUTOTUNER: final configuration " + confExpandedAsStr + " did not compile.\nAbort.") sys.exit(1) # Dump explored configurations as json self.writeJson("exploredConfs.json", self.exploredConfs) return if __name__ == '__main__': argparser = opentuner.default_argparser() autotuneProgram.main(argparser.parse_args())
8,849
29.308219
111
py
noelle
noelle-master/external/autotuner/scripts/filter.py
#!/usr/bin/env python import os import sys import json import traceback thisPath = os.path.dirname(os.path.abspath(__file__)) sys.path.append(thisPath + "/../utils") import utils class Filter: spaceFile = None confFile = None executionTimeFile = None baselineTimeFile = None baselineTime = None ranges = None def getArgs(self): self.spaceFile = os.environ['autotunerSPACE_FILE'] self.confFile = os.environ['INDEX_FILE'] self.executionTimeFile = os.environ['autotunerEXECUTION_TIME'] self.ranges, _ = utils.readSpaceFile(self.spaceFile) self.baselineTimeFile = os.environ['autotunerBASELINE_TIME'] self.baselineTime = utils.readExecutionTimeFile(self.baselineTimeFile) return def writeSpaceFile(self, loopIDsToKeep): space = {} for loopID in self.ranges: if (loopID in loopIDsToKeep): space[loopID] = self.ranges[loopID] else: space[loopID] = [0, 0, 0, 0, 0, 0, 0, 0, 0] utils.writeConfFile(self.spaceFile, space) return def filter(self): techniqueIndex = 3 loopIDsToKeep = [] for loopID in self.ranges: conf = {} testingConf = False for loopIDToZeroOut in self.ranges: conf[loopIDToZeroOut] = [0, 0, 0, 0, 0, 0, 0, 0, 0] if (loopID == loopIDToZeroOut): if (self.ranges[loopIDToZeroOut][techniqueIndex] == 1): # Only DOALL enabled conf[loopIDToZeroOut] = [1, 0, 0, 4, 2, 8, 0, 0, 0] testingConf = True if (not testingConf): loopIDsToKeep.append(loopID) continue # Compile compileRetCode = utils.myCompile(self.confFile, conf) if (compileRetCode != 0): sys.exit(1) # Run parallel optimized binary maxExecutionTime = 2*self.baselineTime runRetCode = utils.myRun(maxExecutionTime) if (runRetCode != 0): sys.exit(1) # Get execution time time = utils.readExecutionTimeFile(self.executionTimeFile) tolerance = 1.2 # 20% if (time < (tolerance*self.baselineTime)): loopIDsToKeep.append(loopID) return loopIDsToKeep if __name__ == '__main__': filterLoops = Filter() filterLoops.getArgs() loopIDsToKeep = filterLoops.filter() filterLoops.writeSpaceFile(loopIDsToKeep)
2,279
24.333333
88
py
noelle
noelle-master/external/autotuner/utils/compileAndRunBaseline.py
#!/usr/bin/env python import os import sys thisPath = os.path.dirname(os.path.abspath(__file__)) sys.path.append(thisPath) import utils def writeExecutionTimeFile(pathToFile, time): with open(str(pathToFile), 'w') as f: f.write(str(time)) f.close() return def compileAndRunBaseline(): spaceFile = os.environ['autotunerSPACE_FILE'] confFile = os.environ['INDEX_FILE'] executionTimeFile = os.environ['autotunerEXECUTION_TIME'] ranges, _ = utils.readSpaceFile(spaceFile) conf = {} for loopID in ranges: conf[loopID] = [] for elem in ranges[loopID]: conf[loopID].append(0) # Compile compileRetCode = utils.myCompile(confFile, conf) if (compileRetCode != 0): sys.exit(1) # Run parallel optimized binary runRetCode = utils.myRun() if (runRetCode != 0): sys.exit(1) # Get execution time time = utils.readExecutionTimeFile(executionTimeFile) return time if __name__ == '__main__': baselineTime = compileAndRunBaseline() baselineTimeFile = os.environ['autotunerBASELINE_TIME'] writeExecutionTimeFile(baselineTimeFile, baselineTime)
1,109
19.943396
59
py
noelle
noelle-master/external/autotuner/utils/utils.py
#!/usr/bin/env python import os import sys from enum import Enum # Risky. It works because autotuner.py and filter.py are at the same level in the directoy tree. thisPath = os.path.dirname(os.path.abspath(__file__)) class Technique(Enum): DOALL = 4 HELIX = 5 DSWP = 6 def readExecutionTimeFile(pathToFile): lineAsFloat = None with open(str(pathToFile), 'r') as f: line = f.readline() lineAsFloat = float(line) f.close() return lineAsFloat def readSpaceFile(pathToFile): ranges = {} loopIDs = [] with open(str(pathToFile), 'r') as f: for line in f.readlines(): loopID = int(line.split()[0]) loopIDs.append(loopID) ranges[loopID] = [] for elem in line.split()[1:]: ranges[loopID].append(elem) f.close() return ranges, loopIDs def writeConfFile(pathToFile, conf): strToWrite = "" for loopID in conf: strToWrite += str(loopID) for elem in conf[loopID]: strToWrite += " " + str(elem) strToWrite += "\n" with open(str(pathToFile), 'w') as f: f.write(strToWrite) f.close() return def myCompile(confFile, conf): # Write autotuner_conf.info file writeConfFile(confFile, conf) return os.system(thisPath + "/../scripts/compile") def myRun(maxExecutionTime = 0): retcode = 0 if (maxExecutionTime == 0): retcode = os.system(thisPath + "/../scripts/run") else: retcode = os.system("timeout " + str(maxExecutionTime) + "s" + " " + thisPath + "/../scripts/run") return retcode
1,509
19.972222
102
py
noelle
noelle-master/external/autotuner/utils/genSeedConf.py
import os import sys import json import tempfile thisPath = os.path.dirname(os.path.abspath(__file__)) sys.path.append(thisPath + "/../utils") import utils # DOALL index is 4 reverseIndex = 4 # We should consider all permutations, but they can grow quickly, and we can't spend all autotuner time running seed configurations. #reverseTechniqueIndexConverter = [[utils.Technique.DOALL-reverseIndex,utils.Technique.HELIX-reverseIndex,utils.Technique.DSWP-reverseIndex], [utils.Technique.DOALL-reverseIndex,utils.Technique.HELIX-reverseIndex], [utils.Technique.DOALL-reverseIndex,utils.Technique.DSWP-reverseIndex], [utils.Technique.HELIX-reverseIndex,utils.Technique.DSWP-reverseIndex], [utils.Technique.DOALL-reverseIndex], [utils.Technique.HELIX-reverseIndex], [utils.Technique.DSWP-reverseIndex]] # So here's an heuristic: if we have multiple choices we tend to pick DOALL reverseTechniqueIndexConverter = [utils.Technique.DOALL-reverseIndex, utils.Technique.DOALL-reverseIndex, utils.Technique.DOALL-reverseIndex, utils.Technique.HELIX-reverseIndex, utils.Technique.DOALL-reverseIndex, utils.Technique.HELIX-reverseIndex, utils.Technique.DSWP-reverseIndex] def readJson(pathToFile): data = {} with open(str(pathToFile)) as f: data = json.load(f) f.close() return data def writeJson(pathToFile, jsonData): with open(pathToFile, 'w') as f: json.dump(jsonData, f) f.close() return def readSeedConf(pathToFile): techniqueToDisableIndex = 3 seedConf = {} with open(str(pathToFile), 'r') as f: i = 0 for line in f: lineAsList = line.split() sys.stderr.write(str(lineAsList) + "\n") # If loop is disabled, then skip the remaining parameters if (int(lineAsList[1]) == 0): i += len(lineAsList[1:]) continue # Skip the loop ID for the seed configuration for elem in lineAsList[1:]: value = int(elem) if (((i - techniqueToDisableIndex) % 9) == 0): seedConf[i] = reverseTechniqueIndexConverter[value] else: if (value != 0): seedConf[i] = value i += 1 f.close() sys.stderr.write(str(seedConf) + "\n") return seedConf def genSeedConfFile(seedConfJson): newfile, pathToFile = tempfile.mkstemp(suffix = '.json') writeJson(pathToFile, seedConfJson) return pathToFile def genSeedConf(pathToSeedConf): seedConfJson = readSeedConf(pathToSeedConf) pathToFile = genSeedConfFile(seedConfJson) return pathToFile if __name__ == '__main__': pathToSeedConf = genSeedConf(sys.argv[1]) print(pathToSeedConf)
2,592
31.012346
471
py
noelle
noelle-master/tests/condor/scripts/generateCondorScript.py
import os import sys import getpass ## Process the command line arguments # # def getArgs(): numOfArgs = len(sys.argv) args = {} args['fromFile'] = str(sys.argv[1]) args['toFile'] = str(sys.argv[2]) args['testsPath'] = str(sys.argv[3]) args['noelleOptions'] = str(sys.argv[4]) args['parOptions'] = str(sys.argv[5]) args['FrontEndOptions'] = str(sys.argv[6]) args['PreMiddleEndOptions'] = str(sys.argv[7]) args['ToolOptions'] = str(sys.argv[8]) args['email'] = str(getpass.getuser()) + '@eecs.northwestern.edu' if (numOfArgs > 9): args['email'] = str(sys.argv[9]) repoPath = '' for elem in str(os.path.dirname(os.path.abspath(__file__))).split(os.sep)[1:-2]: repoPath += os.sep + elem args['repoPath'] = repoPath return args def getNewFile(args): newFileAsStr = '' with open(args['fromFile'], 'r') as f: for line in f: if (line.startswith('Notify_User')): newFileAsStr += 'Notify_User = ' + args['email'] + '\n' elif (line.startswith('RepoPath')): newFileAsStr += 'RepoPath = ' + args['repoPath'] + '\n' elif (line.startswith('TestsPath')): newFileAsStr += 'TestsPath = ' + args['testsPath'] + '\n' elif (line.startswith('ParallelizationOptions')): newFileAsStr += 'ParallelizationOptions = ' + args['parOptions'] + '\n' elif (line.startswith('NoelleOptions')): newFileAsStr += 'NoelleOptions = ' + args['noelleOptions'] + '\n' elif (line.startswith('FrontEndOptions')): newFileAsStr += 'FrontEndOptions = ' + args['FrontEndOptions'] + '\n' elif (line.startswith('PreMiddleEndOptions')): newFileAsStr += 'PreMiddleEndOptions = ' + args['PreMiddleEndOptions'] + '\n' elif (line.startswith('ToolOptions')): newFileAsStr += 'ToolOptions = ' + args['ToolOptions'] + '\n' else: newFileAsStr += str(line) return newFileAsStr def setNewFile(pathToFile, data): with open(str(pathToFile), 'w') as f: f.write(str(data)) f.close() return args = getArgs() newFileAsStr = getNewFile(args) setNewFile(args['toFile'], newFileAsStr)
2,115
30.117647
85
py
ppgn
ppgn-master/settings.py
# Set this to the path to Caffe installation on your system caffe_root = "/path/to/your/caffe/python" gpu = True # ------------------------------------- # The following are hard-coded and hardly change unless we change to use a different generator. # ------------------------------------- # Generator G generator_weights = "nets/generator/noiseless/generator.caffemodel" generator_definition = "nets/generator/noiseless/generator.prototxt" # input / output layers in the generator prototxt generator_in_layer = "feat" generator_out_layer = "deconv0" # Encoder E encoder_weights = "nets/caffenet/bvlc_reference_caffenet.caffemodel" encoder_definition = "nets/caffenet/caffenet.prototxt" # Text files synset_file = "misc/synset_words.txt" vocab_file = "misc/vocabulary.txt"
777
32.826087
95
py
ppgn
ppgn-master/sampling_caption.py
#!/usr/bin/env python ''' Anh Nguyen <anh.ng8@gmail.com> 2016 ''' import os, sys os.environ['GLOG_minloglevel'] = '2' # suprress Caffe verbose prints import settings sys.path.insert(0, settings.caffe_root) import caffe import numpy as np from numpy.linalg import norm import scipy.misc, scipy.io import argparse import util from sampler import Sampler if settings.gpu: caffe.set_mode_gpu() # sampling on GPU (recommended for speed) class CaptionConditionalSampler(Sampler): def __init__ (self, lstm_definition, lstm_weights): self.lstm = caffe.Net(lstm_definition, lstm_weights, caffe.TEST) def forward_backward_from_x_to_condition(self, net, end, image, condition): ''' Forward and backward passes through 'net', the condition model p(y|x), here an image classifier. ''' src = net.blobs['data'] # input image dst = net.blobs[end] sentence = condition['sentence'] previous_word = 0 lstm_layer = "log_prob" feature_layer = "image_features" grad_sum = np.zeros_like(self.lstm.blobs[feature_layer].data) probs = [] for idx, word in enumerate(sentence): if idx > 0: previous_word = sentence[idx - 1] # preparing lstm feature vectors cont = 0 if previous_word == 0 else 1 cont_input = np.array([cont]) word_input = np.array([previous_word]) # Previous word == 0 : meaning this is the start of the sentence # 1. Get feature descriptors from fc8 net.forward(data=image, end=end) descriptor = net.blobs[end].data # 2. Pass this to lstm image_features = np.zeros_like(self.lstm.blobs[feature_layer].data) image_features[:] = descriptor self.lstm.forward(image_features=image_features, cont_sentence=cont_input, input_sentence=word_input, end=lstm_layer) # Display the prediction probs.append ( self.lstm.blobs["probs"].data[0,idx, word] ) self.lstm.blobs[lstm_layer].diff[:, :, word] = 1 diffs = self.lstm.backward(start=lstm_layer, diffs=[feature_layer]) g_word = diffs[feature_layer] # (1000,) grad_sum += g_word # accumulate the gradient from all words # reset objective after each step self.lstm.blobs[lstm_layer].diff.fill(0.) # Average softmax probabilities of all words obj_prob = np.mean(probs) # Backpropagate the gradient from LSTM to the feature extractor convnet dst.diff[...] = grad_sum[0] net.backward(start=end) g = src.diff.copy() dst.diff.fill(0.) # reset objective after each step # Info to be printed out in the below 'print_progress' method info = { } return g, obj_prob, info def get_label(self, condition): return None def print_progress(self, i, info, condition, prob, grad): print "step: %04d\t %s [%.2f]\t norm: [%.2f]" % ( i, condition['readable'], prob, norm(grad) ) def main(): parser = argparse.ArgumentParser(description='Process some integers.') parser.add_argument('--sentence', metavar='w', type=str, default="", nargs='?', help='Sentence to condition on') parser.add_argument('--n_iters', metavar='iter', type=int, default=10, help='Number of sampling steps per each unit') parser.add_argument('--threshold', metavar='w', type=float, default=-1.0, nargs='?', help='The probability threshold to decide whether to keep an image') parser.add_argument('--save_every', metavar='save_iter', type=int, default=1, help='Save a sample every N iterations. 0 to disable saving') parser.add_argument('--reset_every', metavar='reset_iter', type=int, default=0, help='Reset the code every N iterations') parser.add_argument('--lr', metavar='lr', type=float, default=2.0, nargs='?', help='Learning rate') parser.add_argument('--lr_end', metavar='lr', type=float, default=-1.0, nargs='?', help='Ending Learning rate') parser.add_argument('--epsilon2', metavar='lr', type=float, default=1.0, nargs='?', help='Ending Learning rate') parser.add_argument('--epsilon1', metavar='lr', type=float, default=1.0, nargs='?', help='Ending Learning rate') parser.add_argument('--epsilon3', metavar='lr', type=float, default=1.0, nargs='?', help='Ending Learning rate') parser.add_argument('--seed', metavar='n', type=int, default=0, nargs='?', help='Random seed') parser.add_argument('--xy', metavar='n', type=int, default=0, nargs='?', help='Spatial position for conv units') parser.add_argument('--opt_layer', metavar='s', type=str, help='Layer at which we optimize a code') parser.add_argument('--act_layer', metavar='s', type=str, default="fc8", help='Layer at which we activate a neuron') parser.add_argument('--init_file', metavar='s', type=str, default="None", help='Init image') parser.add_argument('--write_labels', action='store_true', default=False, help='Write class labels to images') parser.add_argument('--output_dir', metavar='b', type=str, default=".", help='Output directory for saving results') parser.add_argument('--net_weights', metavar='b', type=str, default=settings.encoder_weights, help='Weights of the net being visualized') parser.add_argument('--net_definition', metavar='b', type=str, default=settings.encoder_definition, help='Definition of the net being visualized') parser.add_argument('--captioner_definition', metavar='b', type=str, help='Definition of the net being visualized') args = parser.parse_args() # Default to constant learning rate if args.lr_end < 0: args.lr_end = args.lr # summary print "-------------" print " sentence: %s" % args.sentence print " n_iters: %s" % args.n_iters print " reset_every: %s" % args.reset_every print " save_every: %s" % args.save_every print " threshold: %s" % args.threshold print " epsilon1: %s" % args.epsilon1 print " epsilon2: %s" % args.epsilon2 print " epsilon3: %s" % args.epsilon3 print " start learning rate: %s" % args.lr print " end learning rate: %s" % args.lr_end print " seed: %s" % args.seed print " opt_layer: %s" % args.opt_layer print " act_layer: %s" % args.act_layer print " init_file: %s" % args.init_file print "-------------" print " output dir: %s" % args.output_dir print " net weights: %s" % args.net_weights print " net definition: %s" % args.net_definition print " captioner definition: %s" % args.captioner_definition print "-------------" # encoder and generator for images encoder = caffe.Net(settings.encoder_definition, settings.encoder_weights, caffe.TEST) generator = caffe.Net(settings.generator_definition, settings.generator_weights, caffe.TEST) # condition network, here an image classification net # this LRCN image captioning net has 1 binary weights but 2 definitions: 1 for feature extractor (AlexNet), 1 for LSTM net = caffe.Net(args.net_definition, args.net_weights, caffe.TEST) # Fix the seed np.random.seed(args.seed) if args.init_file != "None": start_code, start_image = get_code(encoder=encoder, path=args.init_file, layer=args.opt_layer) print "Loaded start code: ", start_code.shape else: # shape of the code being optimized shape = generator.blobs[settings.generator_in_layer].data.shape start_code = np.random.normal(0, 1, shape) # Split the sentence into words words = args.sentence.split("_") sentence = util.convert_words_into_numbers(settings.vocab_file, words) # Condition here is the sentence conditions = [ { "sentence": sentence, "readable": args.sentence.replace("_", " ")} ] # Optimize a code via gradient ascent sampler = CaptionConditionalSampler(args.captioner_definition, args.net_weights) output_image, list_samples = sampler.sampling( condition_net=net, image_encoder=encoder, image_generator=generator, gen_in_layer=settings.generator_in_layer, gen_out_layer=settings.generator_out_layer, start_code=start_code, n_iters=args.n_iters, lr=args.lr, lr_end=args.lr_end, threshold=args.threshold, layer=args.act_layer, conditions=conditions, epsilon1=args.epsilon1, epsilon2=args.epsilon2, epsilon3=args.epsilon3, output_dir=args.output_dir, reset_every=args.reset_every, save_every=args.save_every) # Output image filename = "%s/%s_%04d_%s_h_%s_%s_%s__%s.jpg" % ( args.output_dir, args.act_layer, args.n_iters, args.lr, str(args.epsilon1), str(args.epsilon2), str(args.epsilon3), args.seed ) # Save the final image util.save_image(output_image, filename) print "%s/%s" % (os.getcwd(), filename) # Write labels to images print "Saving images..." for p in list_samples: img, name, label = p util.save_image(img, name) if args.write_labels: util.write_label_to_img(name, label) if __name__ == '__main__': main()
9,371
41.6
157
py
ppgn
ppgn-master/sampler.py
#!/usr/bin/env python ''' Anh Nguyen <anh.ng8@gmail.com> 2017 ''' import os, sys os.environ['GLOG_minloglevel'] = '2' # suprress Caffe verbose prints import settings sys.path.insert(0, settings.caffe_root) import caffe import numpy as np from numpy.linalg import norm import scipy.misc, scipy.io import util class Sampler(object): def backward_from_x_to_h(self, generator, diff, start, end): ''' Backpropagate the gradient from the image (start) back to the latent space (end) of the generator network. ''' dst = generator.blobs[end] dst.diff[...] = diff generator.backward(start=end) g = generator.blobs[start].diff.copy() dst.diff.fill(0.) # reset objective after each step return g def h_autoencoder_grad(self, h, encoder, decoder, gen_out_layer, topleft, inpainting): ''' Compute the gradient of the energy of P(input) wrt input, which is given by decode(encode(input))-input {see Alain & Bengio, 2014}. Specifically, we compute E(G(h)) - h. Note: this is an "upside down" auto-encoder for h that goes h -> x -> h with G modeling h -> x and E modeling x -> h. ''' generated = encoder.forward(feat=h) x = encoder.blobs[gen_out_layer].data.copy() # 256x256 # Crop from 256x256 to 227x227 image_size = decoder.blobs['data'].shape # (1, 3, 227, 227) cropped_x = x[:,:,topleft[0]:topleft[0]+image_size[2], topleft[1]:topleft[1]+image_size[3]] # Mask the image when inpainting if inpainting is not None: cropped_x = util.apply_mask(img=cropped_x, mask=inpainting['mask'], context=inpainting['image']) # Push this 227x227 image through net decoder.forward(data=cropped_x) code = decoder.blobs['fc6'].data g = code - h return g def sampling( self, condition_net, image_encoder, image_generator, gen_in_layer, gen_out_layer, start_code, n_iters, lr, lr_end, threshold, layer, conditions, #units=None, xy=0, epsilon1=1, epsilon2=1, epsilon3=1e-10, inpainting=None, # in-painting args output_dir=None, reset_every=0, save_every=1): # Get the input and output sizes image_shape = condition_net.blobs['data'].data.shape generator_output_shape = image_generator.blobs[gen_out_layer].data.shape encoder_input_shape = image_encoder.blobs['data'].data.shape # Calculate the difference between the input image of the condition net # and the output image from the generator image_size = util.get_image_size(image_shape) generator_output_size = util.get_image_size(generator_output_shape) encoder_input_size = util.get_image_size(encoder_input_shape) # The top left offset to crop the output image to get a 227x227 image topleft = util.compute_topleft(image_size, generator_output_size) topleft_DAE = util.compute_topleft(encoder_input_size, generator_output_size) src = image_generator.blobs[gen_in_layer] # the input feature layer of the generator # Make sure the layer size and initial vector size match assert src.data.shape == start_code.shape # Variables to store the best sample last_xx = np.zeros(image_shape) # best image last_prob = -sys.maxint # highest probability h = start_code.copy() condition_idx = 0 list_samples = [] i = 0 while True: step_size = lr + ((lr_end - lr) * i) / n_iters condition = conditions[condition_idx] # Select a class # 1. Compute the epsilon1 term --- # compute gradient d log(p(h)) / dh per DAE results in Alain & Bengio 2014 d_prior = self.h_autoencoder_grad(h=h, encoder=image_generator, decoder=image_encoder, gen_out_layer=gen_out_layer, topleft=topleft_DAE, inpainting=inpainting) # 2. Compute the epsilon2 term --- # Push the code through the generator to get an image x image_generator.blobs["feat"].data[:] = h generated = image_generator.forward() x = generated[gen_out_layer].copy() # 256x256 # Crop from 256x256 to 227x227 cropped_x = x[:,:,topleft[0]:topleft[0]+image_size[0], topleft[1]:topleft[1]+image_size[1]] cropped_x_copy = cropped_x.copy() if inpainting is not None: cropped_x = util.apply_mask(img=cropped_x, mask=inpainting['mask'], context=inpainting['image']) # Forward pass the image x to the condition net up to an unit k at the given layer # Backprop the gradient through the condition net to the image layer to get a gradient image d_condition_x, prob, info = self.forward_backward_from_x_to_condition(net=condition_net, end=layer, image=cropped_x, condition=condition) if inpainting is not None: # Mask out the class gradient image d_condition_x[:] *= inpainting["mask"] # An additional objective for matching the context image d_context_x256 = np.zeros_like(x.copy()) d_context_x256[:,:,topleft[0]:topleft[0]+image_size[0], topleft[1]:topleft[1]+image_size[1]] = (inpainting["image"] - cropped_x_copy) * inpainting["mask_neg"] d_context_h = self.backward_from_x_to_h(generator=image_generator, diff=d_context_x256, start=gen_in_layer, end=gen_out_layer) # Put the gradient back in the 256x256 format d_condition_x256 = np.zeros_like(x) d_condition_x256[:,:,topleft[0]:topleft[0]+image_size[0], topleft[1]:topleft[1]+image_size[1]] = d_condition_x.copy() # Backpropagate the above gradient all the way to h (through generator) # This gradient 'd_condition' is d log(p(y|h)) / dh (the epsilon2 term in Eq. 11 in the paper) d_condition = self.backward_from_x_to_h(generator=image_generator, diff=d_condition_x256, start=gen_in_layer, end=gen_out_layer) self.print_progress(i, info, condition, prob, d_condition) # 3. Compute the epsilon3 term --- noise = np.zeros_like(h) if epsilon3 > 0: noise = np.random.normal(0, epsilon3, h.shape) # Gaussian noise # Update h according to Eq.11 in the paper d_h = epsilon1 * d_prior + epsilon2 * d_condition + noise # Plus the optional epsilon4 for matching the context region when in-painting if inpainting is not None: d_h += inpainting["epsilon4"] * d_context_h h += step_size/np.abs(d_h).mean() * d_h h = np.clip(h, a_min=0, a_max=30) # Keep the code within a realistic range # Reset the code every N iters (for diversity when running a long sampling chain) if reset_every > 0 and i % reset_every == 0 and i > 0: h = np.random.normal(0, 1, h.shape) # Experimental: For sample diversity, it's a good idea to randomly pick epsilon1 as well epsilon1 = np.random.uniform(low=1e-6, high=1e-2) # Save every sample last_xx = cropped_x.copy() last_prob = prob # Filter samples based on threshold or every N iterations if save_every > 0 and i % save_every == 0 and prob > threshold: name = "%s/samples/%05d.jpg" % (output_dir, i) label = self.get_label(condition) list_samples.append( (last_xx.copy(), name, label) ) # Stop if grad is 0 if norm(d_h) == 0: print " d_h is 0" break # Randomly sample a class every N iterations if i > 0 and i % n_iters == 0: condition_idx += 1 if condition_idx == len(conditions): break i += 1 # Next iter # returning the last sample print "-------------------------" print "Last sample: prob [%s] " % last_prob return last_xx, list_samples
8,319
41.020202
174
py
ppgn
ppgn-master/sampling_class.py
#!/usr/bin/env python ''' Anh Nguyen <anh.ng8@gmail.com> 2016 ''' import os, sys os.environ['GLOG_minloglevel'] = '2' # suprress Caffe verbose prints import settings sys.path.insert(0, settings.caffe_root) import caffe import numpy as np from numpy.linalg import norm import scipy.misc, scipy.io import argparse import util from sampler import Sampler if settings.gpu: caffe.set_mode_gpu() # sampling on GPU class ClassConditionalSampler(Sampler): def __init__ (self): # Load the list of class names with open(settings.synset_file, 'r') as synset_file: self.class_names = [ line.split(",")[0].split(" ", 1)[1].rstrip('\n') for line in synset_file.readlines()] # Hard-coded list of layers that has been tested self.fc_layers = ["fc6", "fc7", "fc8", "loss3/classifier", "fc1000", "prob"] self.conv_layers = ["conv1", "conv2", "conv3", "conv4", "conv5"] def forward_backward_from_x_to_condition(self, net, end, image, condition): ''' Forward and backward passes through 'net', the condition model p(y|x), here an image classifier. ''' unit = condition['unit'] xy = condition['xy'] dst = net.blobs[end] acts = net.forward(data=image, end=end) one_hot = np.zeros_like(dst.data) # Get the activations if end in self.fc_layers: layer_acts = acts[end][0] elif end in self.conv_layers: layer_acts = acts[end][0, :, xy, xy] best_unit = layer_acts.argmax() # highest probability unit # Compute the softmax probs by hand because it's handy in case we want to condition on hidden units as well exp_acts = np.exp(layer_acts - np.max(layer_acts)) probs = exp_acts / (1e-10 + np.sum(exp_acts, keepdims=True)) # The gradient of log of softmax, log(p(y|x)), reduces to: softmax_grad = 1 - probs.copy() obj_prob = probs.flat[unit] # Assign the gradient if end in self.fc_layers: one_hot.flat[unit] = softmax_grad[unit] elif end in self.conv_layers: one_hot[:, unit, xy, xy] = softmax_grad[unit] else: raise Exception("Invalid layer type!") dst.diff[:] = one_hot # Backpropagate the gradient to the image layer diffs = net.backward(start=end, diffs=['data']) g = diffs['data'].copy() dst.diff.fill(0.) # reset objective after each step # Info to be printed out in the below 'print_progress' method info = { 'best_unit': best_unit, 'best_unit_prob': probs.flat[best_unit] } return g, obj_prob, info def get_label(self, condition): unit = condition['unit'] return self.class_names[unit] def print_progress(self, i, info, condition, prob, grad): print "step: %04d\t max: %4s [%.2f]\t obj: %4s [%.2f]\t norm: [%.2f]" % ( i, info['best_unit'], info['best_unit_prob'], condition['unit'], prob, norm(grad) ) def get_code(encoder, path, layer, mask=None): ''' Push the given image through an encoder (here, AlexNet) to get a code. ''' # set up the inputs for the net: image_size = encoder.blobs['data'].shape[2:] # (1, 3, 227, 227) images = np.zeros_like(encoder.blobs["data"].data, dtype='float32') in_image = scipy.misc.imread(path) in_image = scipy.misc.imresize(in_image, (image_size[0], image_size[1])) images[0] = np.transpose(in_image, (2, 0, 1)) # convert to (3, 227, 227) format data = images[:,::-1] # convert from RGB to BGR # subtract the ImageNet mean image_mean = scipy.io.loadmat('misc/ilsvrc_2012_mean.mat')['image_mean'] # (256, 256, 3) topleft = util.compute_topleft(image_size, image_mean.shape[:2]) image_mean = image_mean[topleft[0]:topleft[0]+image_size[0], topleft[1]:topleft[1]+image_size[1]] # crop the image mean data -= np.expand_dims(np.transpose(image_mean, (2,0,1)), 0) # mean is already BGR if mask is not None: data *= mask # initialize the encoder encoder = caffe.Net(settings.encoder_definition, settings.encoder_weights, caffe.TEST) # extract the features encoder.forward(data=data) features = encoder.blobs[layer].data.copy() return features, data def main(): parser = argparse.ArgumentParser(description='Process some integers.') parser.add_argument('--units', metavar='units', type=str, help='an unit to visualize e.g. [0, 999]') parser.add_argument('--n_iters', metavar='iter', type=int, default=10, help='Number of sampling steps per each unit') parser.add_argument('--threshold', metavar='w', type=float, default=-1.0, nargs='?', help='The probability threshold to decide whether to keep an image') parser.add_argument('--save_every', metavar='save_iter', type=int, default=1, help='Save a sample every N iterations. 0 to disable saving') parser.add_argument('--reset_every', metavar='reset_iter', type=int, default=0, help='Reset the code every N iterations') parser.add_argument('--lr', metavar='lr', type=float, default=2.0, nargs='?', help='Learning rate') parser.add_argument('--lr_end', metavar='lr', type=float, default=-1.0, nargs='?', help='Ending Learning rate') parser.add_argument('--epsilon1', metavar='lr', type=float, default=1.0, nargs='?', help='Prior') parser.add_argument('--epsilon2', metavar='lr', type=float, default=1.0, nargs='?', help='Condition') parser.add_argument('--epsilon3', metavar='lr', type=float, default=1.0, nargs='?', help='Noise') parser.add_argument('--epsilon4', metavar='lr', type=float, default=0.0, nargs='?', help='Context') parser.add_argument('--seed', metavar='n', type=int, default=0, nargs='?', help='Random seed') parser.add_argument('--xy', metavar='n', type=int, default=0, nargs='?', help='Spatial position for conv units') parser.add_argument('--opt_layer', metavar='s', type=str, help='Layer at which we optimize a code') parser.add_argument('--act_layer', metavar='s', type=str, default="fc8", help='Layer at which we activate a neuron') parser.add_argument('--init_file', metavar='s', type=str, default="None", help='Init image') parser.add_argument('--write_labels', action='store_true', default=False, help='Write class labels to images') parser.add_argument('--output_dir', metavar='b', type=str, default=".", help='Output directory for saving results') parser.add_argument('--net_weights', metavar='b', type=str, default=settings.encoder_weights, help='Weights of the net being visualized') parser.add_argument('--net_definition', metavar='b', type=str, default=settings.encoder_definition, help='Definition of the net being visualized') args = parser.parse_args() # Default to constant learning rate if args.lr_end < 0: args.lr_end = args.lr # summary print "-------------" print " units: %s xy: %s" % (args.units, args.xy) print " n_iters: %s" % args.n_iters print " reset_every: %s" % args.reset_every print " save_every: %s" % args.save_every print " threshold: %s" % args.threshold print " epsilon1: %s" % args.epsilon1 print " epsilon2: %s" % args.epsilon2 print " epsilon3: %s" % args.epsilon3 print " epsilon4: %s" % args.epsilon4 print " start learning rate: %s" % args.lr print " end learning rate: %s" % args.lr_end print " seed: %s" % args.seed print " opt_layer: %s" % args.opt_layer print " act_layer: %s" % args.act_layer print " init_file: %s" % args.init_file print "-------------" print " output dir: %s" % args.output_dir print " net weights: %s" % args.net_weights print " net definition: %s" % args.net_definition print "-------------" # encoder and generator for images encoder = caffe.Net(settings.encoder_definition, settings.encoder_weights, caffe.TEST) generator = caffe.Net(settings.generator_definition, settings.generator_weights, caffe.TEST) # condition network, here an image classification net net = caffe.Classifier(args.net_definition, args.net_weights, mean = np.float32([104.0, 117.0, 123.0]), # ImageNet mean channel_swap = (2,1,0)) # the reference model has channels in BGR order instead of RGB # Fix the seed np.random.seed(args.seed) # Sampler for class-conditional generation sampler = ClassConditionalSampler() inpainting = None if args.init_file != "None": # Pre-compute masks if we want to perform inpainting if args.epsilon4 > 0: mask, neg = util.get_mask() else: neg = None # Get the code for the masked image start_code, start_image = get_code(encoder=encoder, path=args.init_file, layer=args.opt_layer, mask=neg) # Package settings for in-painting experiments if args.epsilon4 > 0: inpainting = { "mask" : mask, "mask_neg" : neg, "image" : start_image, "epsilon4" : args.epsilon4 } print "Loaded init code: ", start_code.shape else: # shape of the code being optimized shape = generator.blobs[settings.generator_in_layer].data.shape start_code = np.random.normal(0, 1, shape) print ">>", np.min(start_code), np.max(start_code) # Separate the dash-separated list of units into numbers conditions = [ { "unit": int(u), "xy": args.xy } for u in args.units.split("_") ] # Optimize a code via gradient ascent output_image, list_samples = sampler.sampling( condition_net=net, image_encoder=encoder, image_generator=generator, gen_in_layer=settings.generator_in_layer, gen_out_layer=settings.generator_out_layer, start_code=start_code, n_iters=args.n_iters, lr=args.lr, lr_end=args.lr_end, threshold=args.threshold, layer=args.act_layer, conditions=conditions, epsilon1=args.epsilon1, epsilon2=args.epsilon2, epsilon3=args.epsilon3, inpainting=inpainting, output_dir=args.output_dir, reset_every=args.reset_every, save_every=args.save_every) # Output image filename = "%s/%s_%04d_%04d_%s_h_%s_%s_%s_%s__%s.jpg" % ( args.output_dir, args.act_layer, conditions[0]["unit"], args.n_iters, args.lr, str(args.epsilon1), str(args.epsilon2), str(args.epsilon3), str(args.epsilon4), args.seed ) if inpainting != None: output_image = util.stitch(start_image, output_image) # Save the final image util.save_image(output_image, filename) print "%s/%s" % (os.getcwd(), filename) # Write labels to images print "Saving images..." for p in list_samples: img, name, label = p util.save_image(img, name) if args.write_labels: util.write_label_to_img(name, label) if __name__ == '__main__': main()
11,251
40.216117
165
py
ppgn
ppgn-master/util.py
import numpy as np import scipy.misc import subprocess def normalize(img, out_range=(0.,1.), in_range=None): if not in_range: min_val = np.min(img) max_val = np.max(img) else: min_val = in_range[0] max_val = in_range[1] result = np.copy(img) result[result > max_val] = max_val result[result < min_val] = min_val result = (result - min_val) / (max_val - min_val) * (out_range[1] - out_range[0]) + out_range[0] return result def deprocess(images, out_range=(0.,1.), in_range=None): num = images.shape[0] c = images.shape[1] ih = images.shape[2] iw = images.shape[3] result = np.zeros((ih, iw, 3)) # Normalize before saving result[:] = images[0].copy().transpose((1,2,0)) result = normalize(result, out_range, in_range) return result def get_image_size(data_shape): ''' Return (227, 227) from (1, 3, 227, 227) tensor. ''' if len(data_shape) == 4: return data_shape[2:] else: raise Exception("Data shape invalid.") def save_image(img, name): ''' Normalize and save the image. ''' img = img[:,::-1, :, :] # Convert from BGR to RGB output_img = deprocess(img, in_range=(-120,120)) scipy.misc.imsave(name, output_img) def write_label_to_img(filename, label): # Add a label below each image via ImageMagick subprocess.call(["convert %s -gravity south -splice 0x10 %s" % (filename, filename)], shell=True) subprocess.call(["convert %s -append -gravity Center -pointsize %s label:\"%s\" -border 0x0 -append %s" % (filename, 30, label, filename)], shell=True) def convert_words_into_numbers(vocab_file, words): # Load vocabularty f = open(vocab_file, 'r') lines = f.read().splitlines() numbers = [ lines.index(w) + 1 for w in words ] numbers.append( 0 ) # <unk> return numbers def get_mask(): ''' Compute the binary mask to be used for inpainting experiments. ''' image_shape = (3, 227, 227) # Make a blob of noise in the center mask = np.zeros(image_shape) mask_neg = np.ones(image_shape) # width and height of the mask w, h = (100, 100) # starting and ending positions of mask max_x, max_y = image_shape[1] - w, image_shape[2] - h x0 = np.random.randint(low=0, high=max_x) x1 = np.min([ image_shape[1], x0 + w ]) y0 = np.random.randint(low=0, high=max_y) y1 = np.min([ image_shape[2], y0 + h ]) for y in np.arange(x0, x1): for x in np.arange(y0, y1): mask [ :, x, y ] = 1 mask_neg [ :, x, y ] = 0 return mask, mask_neg def compute_topleft(input_size, output_size): ''' Compute the offsets (top, left) to crop the output image if its size does not match that of the input image. The output size is fixed at 256 x 256 as the generator network is trained on 256 x 256 images. However, the input size often changes depending on the network. ''' assert len(input_size) == 2, "input_size must be (h, w)" assert len(output_size) == 2, "output_size must be (h, w)" topleft = ((output_size[0] - input_size[0])/2, (output_size[1] - input_size[1])/2) return topleft def apply_mask(img, mask, context): assert len(img.shape) == 4 assert img.shape[0] == 1 assert img.shape[1] == 3 # Mask out a patch (defined by the binary "mask") img[0] *= mask img += context return img def stitch(left, right): ''' Stitch two images together horizontally. ''' assert len(left.shape) == 4 assert len(right.shape) == 4 assert left.shape[0] == 1 assert right.shape[0] == 1 # Save final image and the masked image image_size = right.shape[2] separator_width = 1 canvas_size = image_size * 2 + separator_width output = np.zeros( (1, 3, image_size, canvas_size) ) output.fill(255.0) output[:,:,:image_size,:image_size] = left output[:,:,:,image_size + separator_width:] = right return output
4,035
27.422535
112
py
gcnn-survey-paper
gcnn-survey-paper-master/best_model.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Averages validation metric over multiple runs and returns best model.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os from absl import app from absl import flags import numpy as np import scipy.stats as stats import tensorflow as tf flags.DEFINE_string('dir', '/tmp/launch', 'path were models are saved.') flags.DEFINE_string('target', 'node_acc', 'target metric to use.') flags.DEFINE_string('datasets', 'cora', 'datasets to use.') flags.DEFINE_string('drop_prop', '0-10-20-30-40-50-60-70-80-90', 'proportion of edges dropped') flags.DEFINE_string('save_file', 'best_params', 'name of files to same the' 'results.') flags.DEFINE_string('models', 'Gcn', 'name of model directories to parse.') FLAGS = flags.FLAGS def get_val_test_acc(data): """Parses log file to retrieve test and val accuracy.""" data = [x.split() for x in data if len(x.split()) > 1] val_acc_idx = data[-4].index('val_{}'.format(FLAGS.target)) test_acc_idx = data[-3].index('test_{}'.format(FLAGS.target)) val_acc = data[-4][val_acc_idx + 2] test_acc = data[-3][test_acc_idx + 2] return float(val_acc) * 100, float(test_acc) * 100 def main(_): log_file = tf.gfile.Open(os.path.join(FLAGS.dir, FLAGS.save_file), 'w') for dataset in FLAGS.datasets.split('-'): for prop in FLAGS.drop_prop.split('-'): dir_path = os.path.join(FLAGS.dir, dataset, prop) if tf.gfile.IsDirectory(dir_path): print(dir_path) for model_name in tf.gfile.ListDirectory(dir_path): if model_name in FLAGS.models.split('-'): model_dir = os.path.join(dir_path, model_name) train_log_files = [ filename for filename in tf.gfile.ListDirectory(model_dir) if 'log' in filename ] eval_stats = {} for filename in train_log_files: data = tf.gfile.Open(os.path.join(model_dir, filename)).readlines() nb_lines = len(data) if nb_lines > 0: if 'Training done' in data[-1]: val_acc, test_acc = get_val_test_acc(data) params = '-'.join(filename.split('-')[:-1]) if params in eval_stats: eval_stats[params]['val'].append(val_acc) eval_stats[params]['test'].append(test_acc) else: eval_stats[params] = {'val': [val_acc], 'test': [test_acc]} best_val_metric = -1 best_params = None for params in eval_stats: val_metric = np.mean(eval_stats[params]['val']) if val_metric > best_val_metric: best_val_metric = val_metric best_params = params # print(eval_stats) log_file.write('\n' + model_dir + '\n') log_file.write('Best params: {}\n'.format(best_params)) log_file.write('val_{}: {} +- {}\n'.format( FLAGS.target, round(np.mean(eval_stats[best_params]['val']), 2), round(stats.sem(eval_stats[best_params]['val']), 2))) log_file.write('test_{}: {} +- {}\n'.format( FLAGS.target, round( np.mean(eval_stats[best_params]['test']), 2), round(stats.sem(eval_stats[best_params]['test']), 2))) if __name__ == '__main__': app.run(main)
4,070
40.540816
80
py
gcnn-survey-paper
gcnn-survey-paper-master/launch.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Train models with different combinations of parameters.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from itertools import product import os from absl import app from absl import flags from train import Config from train import TrainTest flags.DEFINE_string('launch_save_dir', '/tmp/launch', 'Where to save the results.') flags.DEFINE_string('launch_model_name', 'Gcn', 'Model to train.') flags.DEFINE_string('launch_dataset', 'cora', 'Dataset to use.') flags.DEFINE_string('launch_datapath', 'data/', 'Path to data folder.') flags.DEFINE_boolean('launch_sparse_features', True, 'True if node features are sparse.') flags.DEFINE_boolean('launch_normalize_adj', True, 'True to normalize adjacency matrix') flags.DEFINE_integer('launch_n_runs', 5, 'number of runs for each combination of parameters.') FLAGS = flags.FLAGS def get_params(): ############################### CHANGE PARAMS HERE ########################## return { # training parameters 'lr': [0.01], 'epochs': [10000], 'patience': [10], 'node_l2_reg': [0.001, 0.0005], 'edge_l2_reg': [0.], 'edge_reg': [0], 'p_drop_node': [0.5], 'p_drop_edge': [0], # model parameters 'n_hidden_node': ['128', '64'], 'n_att_node': ['8-8'], 'n_hidden_edge': ['128-64'], 'n_att_edge': ['8-1'], 'topk': [0], 'att_mechanism': ['l2'], 'edge_loss': ['w_sigmoid_ce'], 'cheby_k_loc': [1], 'semi_emb_k': [-1], # data parameters 'drop_edge_prop': [0, 50], 'normalize_adj': [True] } ############################################################################# def get_config(run_params, data): """Parse configuration parameters for training.""" config = Config() for param in run_params: if 'n_hidden' in param or 'n_att' in param: # Number of layers and att are defined as string so we parse # them differently setattr(config, param, list(map(int, run_params[param].split('-')))) else: setattr(config, param, run_params[param]) config.set_num_nodes_edges(data) return config def main(_): params = get_params() trainer = TrainTest(FLAGS.launch_model_name) print('Loading dataset...') trainer.load_dataset(FLAGS.launch_dataset, FLAGS.launch_sparse_features, FLAGS.launch_datapath) print('Dataset loaded!') # iterate over all combination of parameters all_params = product(*params.values()) for run_params in all_params: run_params = dict(zip(params, run_params)) # load the dataset and process adjacency and node features trainer.mask_edges(trainer.data['adj_true'], run_params['drop_edge_prop']) trainer.process_adj(FLAGS.launch_normalize_adj) config = get_config(run_params, trainer.data) # multilple runs save_dir = os.path.join(FLAGS.launch_save_dir, FLAGS.launch_dataset, str(run_params['drop_edge_prop']), FLAGS.launch_model_name) for run_id in range(FLAGS.launch_n_runs): filename_suffix = config.get_filename_suffix(run_id) trainer.run(config, save_dir, filename_suffix) if __name__ == '__main__': app.run(main)
3,950
32.769231
79
py
gcnn-survey-paper
gcnn-survey-paper-master/__init__.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License.
565
39.428571
73
py
gcnn-survey-paper
gcnn-survey-paper-master/train.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Training script for GNN models for link prediction/node classification.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import datetime import os from absl import app from absl import flags import models.edge_models as edge_models import models.node_edge_models as node_edge_models import models.node_models as node_models import numpy as np import scipy.sparse as sp import tensorflow as tf from utils.data_utils import load_data from utils.data_utils import mask_test_edges from utils.data_utils import mask_val_test_edges from utils.data_utils import process_adj from utils.data_utils import sparse_to_tuple from utils.train_utils import check_improve from utils.train_utils import format_metrics from utils.train_utils import format_params flags.DEFINE_string('model_name', 'Gat', 'Which model to use.') flags.DEFINE_integer('epochs', 10000, 'Number of epochs to train for.') flags.DEFINE_integer('patience', 100, 'Patience for early stopping.') flags.DEFINE_string('dataset', 'cora', 'Dataset to use: (cora - citeseer - pubmed).') flags.DEFINE_string('datapath', 'data/', 'Path to directory with data files.') flags.DEFINE_string('save_dir', '/tmp/models/cora/gat', 'Directory where to save model checkpoints and summaries.') flags.DEFINE_float('lr', 0.005, 'Learning rate to use.') flags.DEFINE_string( 'model_checkpoint', '', 'Model checkpoint to load before' 'training or for testing. If not specified the model will be trained from ' 'scratch.') flags.DEFINE_float('drop_edge_prop', 0, 'Percentage of edges to remove (0 to keep all edges).') flags.DEFINE_float('node_l2_reg', 0.0005, 'L2 regularization to use for node ' 'model parameters.') flags.DEFINE_float('edge_l2_reg', 0., 'L2 regularization to use for edge ' 'model parameters.') flags.DEFINE_float('edge_reg', 0., 'Regularization to use for the edge ' 'loss.') flags.DEFINE_integer( 'cheby_k_loc', 1, 'K for K-localized filters in Chebyshev' 'polynomials approximation.') flags.DEFINE_integer( 'semi_emb_k', -1, 'which layer to regularize for' 'semi-supervised embedding model.') flags.DEFINE_float('p_drop_node', 0.6, 'Dropout probability for node model.') flags.DEFINE_float('p_drop_edge', 0., 'Dopout probability for edge model.') flags.DEFINE_integer( 'topk', 1000, 'Top k entries to keep in adjacency for' ' NodeEdge models.') flags.DEFINE_string( 'n_hidden_node', '8', 'Number of hidden units per layer in node model. ' 'The last layer has as many nodes as the number of classes ' 'in the dataset.') flags.DEFINE_string( 'n_att_node', '8-1', 'Number of attentions heads per layer in for node model. ' '(This is only for graph attention models).') flags.DEFINE_string('n_hidden_edge', '32-16', 'Number of hidden units per layer in edge model.') flags.DEFINE_string( 'n_att_edge', '8-4', 'Number of attentions heads per layer in for edge model. ' '(This is only for edge graph attention models)') flags.DEFINE_string( 'att_mechanism', 'dot', 'Attention mehcanism to use: dot product, asymmetric dot ' 'product or attention (dot - att - asym-dot).') flags.DEFINE_string( 'edge_loss', 'w_sigmoid_ce', 'edge loss (w_sigmoid_ce - neg_sampling_ce). ' 'w_sigmoid_ce for weighted sigmoid cross entropy and neg_sampling_ce for' 'negative sampling.') flags.DEFINE_boolean('sparse_features', True, 'True if node features are sparse.') flags.DEFINE_boolean( 'normalize_adj', True, 'Whether to normalize adjaceny or not (True for' 'GCN models and False for GAT models).') flags.DEFINE_integer('run_id', 0, 'Run id.') FLAGS = flags.FLAGS NODE_MODELS = ['Gat', 'Gcn', 'Mlp', 'Hgat', 'Pgcn', 'SemiEmb', 'Cheby'] NODE_EDGE_MODELS = [ 'GaeGat', 'GaeGcn', 'GatGraphite', 'GaeGatConcat', 'GaeGcnConcat', 'Gcat' ] EDGE_MODELS = ['Gae', 'Egat', 'Emlp', 'Vgae'] class Config(object): """Gets config parameters from flags to train the GNN models.""" def __init__(self): # Model parameters self.n_hidden_node = list(map(int, FLAGS.n_hidden_node.split('-'))) self.n_att_node = list(map(int, FLAGS.n_att_node.split('-'))) self.n_hidden_edge = list(map(int, FLAGS.n_hidden_edge.split('-'))) self.n_att_edge = list(map(int, FLAGS.n_att_edge.split('-'))) self.topk = FLAGS.topk self.att_mechanism = FLAGS.att_mechanism self.edge_loss = FLAGS.edge_loss self.cheby_k_loc = FLAGS.cheby_k_loc self.semi_emb_k = FLAGS.semi_emb_k # Dataset parameters self.sparse_features = FLAGS.sparse_features # Training parameters self.lr = FLAGS.lr self.epochs = FLAGS.epochs self.patience = FLAGS.patience self.node_l2_reg = FLAGS.node_l2_reg self.edge_l2_reg = FLAGS.edge_l2_reg self.edge_reg = FLAGS.edge_reg self.p_drop_node = FLAGS.p_drop_node self.p_drop_edge = FLAGS.p_drop_edge def set_num_nodes_edges(self, data): if self.sparse_features: self.nb_nodes, self.input_dim = data['features'][-1] else: self.nb_nodes, self.input_dim = data['features'].shape self.nb_classes = data['node_labels'].shape[-1] self.n_hidden_node += [int(self.nb_classes)] self.nb_edges = np.sum(data['adj_train'] > 0) - self.nb_nodes self.multilabel = np.max(np.sum(data['node_labels'], 1)) > 1 def get_filename_suffix(self, run_id): """Formats all params in a string for log file suffix.""" all_params = [ self.lr, self.epochs, self.patience, self.node_l2_reg, self.edge_l2_reg, self.edge_reg, self.p_drop_node, self.p_drop_edge, '.'.join([ str(x) for x in self.n_hidden_node ]), '.'.join([str(x) for x in self.n_att_node]), '.'.join([str(x) for x in self.n_hidden_edge]), '.'.join( [str(x) for x in self.n_att_edge]), self.topk, self.att_mechanism, self.edge_loss, self.cheby_k_loc, self.semi_emb_k, run_id ] file_suffix = '-'.join([str(x) for x in all_params]) return file_suffix class TrainTest(object): """Class to train node and edge classification models""" def __init__(self, model_name): # initialize global step self.global_step = 0 self.model_name = model_name self.data = {'train': {}, 'test': {}, 'val': {}} def load_dataset(self, dataset, sparse_features, datapath): """Loads citation dataset.""" dataset = load_data(dataset, datapath) adj_true = dataset[0] + sp.eye(dataset[0].shape[0]) # adj_true to compute link prediction metrics self.data['adj_true'] = adj_true.todense() if sparse_features: self.data['features'] = sparse_to_tuple(dataset[1]) else: self.data['features'] = dataset[1] self.data['node_labels'] = dataset[2] self.data['train']['node_mask'] = dataset[3] self.data['val']['node_mask'] = dataset[4] self.data['test']['node_mask'] = dataset[5] def mask_edges(self, adj_true, drop_edge_prop): """Load edge mask and remove edges for training adjacency.""" # adj_train to compute loss if drop_edge_prop > 0: if self.model_name in NODE_MODELS: self.data['adj_train'], test_mask = mask_test_edges( sp.coo_matrix(adj_true), drop_edge_prop * 0.01) else: self.data['adj_train'], val_mask, test_mask = mask_val_test_edges( sp.coo_matrix(adj_true), drop_edge_prop * 0.01) self.data['val']['edge_mask'] = val_mask self.data['train']['edge_mask'] = val_mask # unused self.data['test']['edge_mask'] = test_mask self.data['adj_train'] += sp.eye(adj_true.shape[0]) self.data['adj_train'] = self.data['adj_train'].todense() else: self.data['adj_train'] = adj_true def process_adj(self, norm_adj): # adj_train_norm for inference if norm_adj: adj_train_norm = process_adj(self.data['adj_train'], self.model_name) else: adj_train_norm = sp.coo_matrix(self.data['adj_train']) self.data['adj_train_norm'] = sparse_to_tuple(adj_train_norm) def init_global_step(self): self.global_step = 0 def create_saver(self, save_dir, filename_suffix): """Creates saver to save model checkpoints.""" self.summary_writer = tf.summary.FileWriter( save_dir, tf.get_default_graph(), filename_suffix=filename_suffix) self.saver = tf.train.Saver() # logging file to print metrics and loss self.log_file = tf.gfile.Open( os.path.join(save_dir, '{}.log'.format(filename_suffix)), 'w') def _create_summary(self, loss, metrics, split): """Create summaries for tensorboard.""" with tf.name_scope('{}-summary'.format(split)): tf.summary.scalar('loss', loss) for metric in metrics: tf.summary.scalar(metric, metrics[metric]) summary_op = tf.summary.merge_all() return summary_op def _make_feed_dict(self, split): """Creates feed dictionnaries for edge models and node models.""" if split == 'train': is_training = True else: is_training = False return self.model.make_feed_dict(self.data, split, is_training) def _get_model_and_targets(self, multilabel): """Define targets to select best model based on val metrics.""" if self.model_name in NODE_MODELS: model_class = getattr(node_models, self.model_name) if multilabel: target_metrics = {'f1': 1, 'loss': 0} else: target_metrics = {'node_acc': 1, 'loss': 0} # target_metrics = {'node_acc': 1} elif self.model_name in NODE_EDGE_MODELS: model_class = getattr(node_edge_models, self.model_name) target_metrics = {'node_acc': 1} else: model_class = getattr(edge_models, self.model_name) target_metrics = {'edge_pr_auc': 1} #, 'loss': 0} return model_class, target_metrics def build_model(self, config): """Build model graph.""" model_class, self.target_metrics = self._get_model_and_targets( config.multilabel) self.model = model_class(config) all_ops = self.model.build_graph() loss, train_op, metric_op, metric_update_op = all_ops self.train_ops = [train_op] self.eval_ops = [loss, metric_update_op] self.metrics = metric_op self.train_summary = self._create_summary(loss, metric_op, 'train') self.val_summary = self._create_summary(loss, metric_op, 'val') def _eval_model(self, sess, split): """Evaluates model.""" sess.run(tf.local_variables_initializer()) if split == 'train': metrics = {} # tmp way to not eval on train for edge model metrics['loss'] = sess.run( self.eval_ops[0], feed_dict=self._make_feed_dict(split)) else: loss, _ = sess.run(self.eval_ops, feed_dict=self._make_feed_dict(split)) metrics = sess.run(self.metrics, feed_dict=self._make_feed_dict(split)) metrics['loss'] = loss return metrics def _init_best_metrics(self): best_metrics = {} for metric in self.target_metrics: if self.target_metrics[metric] == 1: best_metrics[metric] = -1 else: best_metrics[metric] = np.inf return best_metrics def _log(self, message): """Writes into train.log file.""" time = datetime.datetime.now().strftime('%d.%b %Y %H:%M:%S') self.log_file.write(time + ' : ' + message + '\n') def init_model_weights(self, sess): self._log('Initializing model weights...') sess.run(tf.global_variables_initializer()) sess.run(tf.local_variables_initializer()) def restore_checkpoint(self, sess, model_checkpoint=None): """Loads model checkpoint if found and computes evaluation metrics.""" if model_checkpoint is None or not tf.train.checkpoint_exists( model_checkpoint): self.init_model_weights(sess) else: self._log('Loading existing model saved at {}'.format(model_checkpoint)) self.saver.restore(sess, model_checkpoint) self.global_step = int(model_checkpoint.split('-')[-1]) val_metrics = self._eval_model(sess, 'val') test_metrics = self._eval_model(sess, 'test') self._log(format_metrics(val_metrics, 'val')) self._log(format_metrics(test_metrics, 'test')) def train(self, sess, config): """Trains node classification model or joint node edge model.""" self._log('Training {} model...'.format(self.model_name)) self._log('Training parameters : \n ' + format_params(config)) epochs = config.epochs lr = config.lr patience = config.patience # best_step = self.global_step # step for patience curr_step = 0 # best metrics to select model best_val_metrics = self._init_best_metrics() best_test_metrics = self._init_best_metrics() # train the model for epoch in range(epochs): self.global_step += 1 sess.run(self.train_ops, feed_dict=self._make_feed_dict('train')) train_metrics = self._eval_model(sess, 'train') val_metrics = self._eval_model(sess, 'val') self._log('Epoch {} : lr = {:.4f} | '.format(epoch, lr) + format_metrics(train_metrics, 'train') + format_metrics(val_metrics, 'val')) # write summaries train_summary = sess.run(self.train_summary, self._make_feed_dict('train')) val_summary = sess.run(self.val_summary, self._make_feed_dict('val')) self.summary_writer.add_summary( train_summary, global_step=self.global_step) self.summary_writer.add_summary(val_summary, global_step=self.global_step) # save model checkpoint if val acc increased and val loss decreased comp = check_improve(best_val_metrics, val_metrics, self.target_metrics) if np.any(comp): if np.all(comp): # best_step = self.global_step # save_path = os.path.join(save_dir, 'model') # self.saver.save(sess, save_path, global_step=self.global_step) best_test_metrics = self._eval_model(sess, 'test') best_val_metrics = val_metrics curr_step = 0 else: curr_step += 1 if curr_step == patience: self._log('Early stopping') break self._log('\n' + '*' * 40 + ' Best model metrics ' + '*' * 40) # load best model to evaluate on test set # save_path = os.path.join(save_dir, 'model-{}'.format(best_step)) # self.restore_checkpoint(sess, save_path) self._log(format_metrics(best_val_metrics, 'val')) self._log(format_metrics(best_test_metrics, 'test')) self._log('\n' + '*' * 40 + ' Training done ' + '*' * 40) def run(self, config, save_dir, file_prefix): """Build and train a model.""" tf.reset_default_graph() self.init_global_step() # build model self.build_model(config) # create summary writer and save for model weights if not os.path.exists(save_dir): tf.gfile.MakeDirs(save_dir) self.create_saver(save_dir, file_prefix) # run sessions with tf.Session() as sess: self.init_model_weights(sess) self.train(sess, config) sess.close() self.log_file.close() def main(_): # parse configuration parameters trainer = TrainTest(FLAGS.model_name) print('Loading dataset...') # load the dataset and process adjacency and node features trainer.load_dataset(FLAGS.dataset, FLAGS.sparse_features, FLAGS.datapath) trainer.mask_edges(trainer.data['adj_true'], FLAGS.drop_edge_prop) trainer.process_adj(FLAGS.normalize_adj) print('Dataset loaded...') config = Config() config.set_num_nodes_edges(trainer.data) filename_suffix = config.get_filename_suffix(FLAGS.run_id) trainer.run(config, FLAGS.save_dir, filename_suffix) if __name__ == '__main__': app.run(main)
16,343
38.669903
80
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/__init__.py
0
0
0
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/setup.py
from setuptools import setup from setuptools import find_packages setup(name='gcn', version='1.0', description='Graph Convolutional Networks in Tensorflow', author='Thomas Kipf', author_email='thomas.kipf@gmail.com', url='https://tkipf.github.io', download_url='https://github.com/tkipf/gcn', license='MIT', install_requires=['numpy', 'tensorflow', 'networkx', 'scipy' ], package_data={'gcn': ['README.md']}, packages=find_packages())
591
31.888889
63
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/__init__.py
0
0
0
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/inits.py
import tensorflow as tf import numpy as np def uniform(shape, scale=0.05, name=None): """Uniform init.""" initial = tf.random_uniform(shape, minval=-scale, maxval=scale, dtype=tf.float32) return tf.Variable(initial, name=name) def glorot(shape, name=None): """Glorot & Bengio (AISTATS 2010) init.""" init_range = np.sqrt(6.0/(shape[0]+shape[1])) initial = tf.random_uniform(shape, minval=-init_range, maxval=init_range, dtype=tf.float32) return tf.Variable(initial, name=name) def zeros(shape, name=None): """All zeros.""" initial = tf.zeros(shape, dtype=tf.float32) return tf.Variable(initial, name=name) def ones(shape, name=None): """All ones.""" initial = tf.ones(shape, dtype=tf.float32) return tf.Variable(initial, name=name)
791
28.333333
95
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/utils.py
import numpy as np import pickle as pkl import networkx as nx import scipy.sparse as sp from scipy.sparse.linalg.eigen.arpack import eigsh import sys def parse_index_file(filename): """Parse index file.""" index = [] for line in open(filename): index.append(int(line.strip())) return index def sample_mask(idx, l): """Create mask.""" mask = np.zeros(l) mask[idx] = 1 return np.array(mask, dtype=np.bool) def load_data(dataset_str): """ Loads input data from gcn/data directory ind.dataset_str.x => the feature vectors of the training instances as scipy.sparse.csr.csr_matrix object; ind.dataset_str.tx => the feature vectors of the test instances as scipy.sparse.csr.csr_matrix object; ind.dataset_str.allx => the feature vectors of both labeled and unlabeled training instances (a superset of ind.dataset_str.x) as scipy.sparse.csr.csr_matrix object; ind.dataset_str.y => the one-hot labels of the labeled training instances as numpy.ndarray object; ind.dataset_str.ty => the one-hot labels of the test instances as numpy.ndarray object; ind.dataset_str.ally => the labels for instances in ind.dataset_str.allx as numpy.ndarray object; ind.dataset_str.graph => a dict in the format {index: [index_of_neighbor_nodes]} as collections.defaultdict object; ind.dataset_str.test.index => the indices of test instances in graph, for the inductive setting as list object. All objects above must be saved using python pickle module. :param dataset_str: Dataset name :return: All data input files loaded (as well the training/test data). """ names = ['x', 'y', 'tx', 'ty', 'allx', 'ally', 'graph'] objects = [] for i in range(len(names)): with open("data/ind.{}.{}".format(dataset_str, names[i]), 'rb') as f: if sys.version_info > (3, 0): objects.append(pkl.load(f, encoding='latin1')) else: objects.append(pkl.load(f)) x, y, tx, ty, allx, ally, graph = tuple(objects) test_idx_reorder = parse_index_file("data/ind.{}.test.index".format(dataset_str)) test_idx_range = np.sort(test_idx_reorder) if dataset_str == 'citeseer': # Fix citeseer dataset (there are some isolated nodes in the graph) # Find isolated nodes, add them as zero-vecs into the right position test_idx_range_full = range(min(test_idx_reorder), max(test_idx_reorder)+1) tx_extended = sp.lil_matrix((len(test_idx_range_full), x.shape[1])) tx_extended[test_idx_range-min(test_idx_range), :] = tx tx = tx_extended ty_extended = np.zeros((len(test_idx_range_full), y.shape[1])) ty_extended[test_idx_range-min(test_idx_range), :] = ty ty = ty_extended features = sp.vstack((allx, tx)).tolil() features[test_idx_reorder, :] = features[test_idx_range, :] adj = nx.adjacency_matrix(nx.from_dict_of_lists(graph)) labels = np.vstack((ally, ty)) labels[test_idx_reorder, :] = labels[test_idx_range, :] idx_test = test_idx_range.tolist() idx_train = range(len(y)) idx_val = range(len(y), len(y)+500) train_mask = sample_mask(idx_train, labels.shape[0]) val_mask = sample_mask(idx_val, labels.shape[0]) test_mask = sample_mask(idx_test, labels.shape[0]) y_train = np.zeros(labels.shape) y_val = np.zeros(labels.shape) y_test = np.zeros(labels.shape) y_train[train_mask, :] = labels[train_mask, :] y_val[val_mask, :] = labels[val_mask, :] y_test[test_mask, :] = labels[test_mask, :] return adj, features, y_train, y_val, y_test, train_mask, val_mask, test_mask def sparse_to_tuple(sparse_mx): """Convert sparse matrix to tuple representation.""" def to_tuple(mx): if not sp.isspmatrix_coo(mx): mx = mx.tocoo() coords = np.vstack((mx.row, mx.col)).transpose() values = mx.data shape = mx.shape return coords, values, shape if isinstance(sparse_mx, list): for i in range(len(sparse_mx)): sparse_mx[i] = to_tuple(sparse_mx[i]) else: sparse_mx = to_tuple(sparse_mx) return sparse_mx def preprocess_features(features): """Row-normalize feature matrix and convert to tuple representation""" rowsum = np.array(features.sum(1)) r_inv = np.power(rowsum, -1).flatten() r_inv[np.isinf(r_inv)] = 0. r_mat_inv = sp.diags(r_inv) features = r_mat_inv.dot(features) return sparse_to_tuple(features) def normalize_adj(adj): """Symmetrically normalize adjacency matrix.""" adj = sp.coo_matrix(adj) rowsum = np.array(adj.sum(1)) d_inv_sqrt = np.power(rowsum, -0.5).flatten() d_inv_sqrt[np.isinf(d_inv_sqrt)] = 0. d_mat_inv_sqrt = sp.diags(d_inv_sqrt) return adj.dot(d_mat_inv_sqrt).transpose().dot(d_mat_inv_sqrt).tocoo() def preprocess_adj(adj): """Preprocessing of adjacency matrix for simple GCN model and conversion to tuple representation.""" adj_normalized = normalize_adj(adj + sp.eye(adj.shape[0])) return sparse_to_tuple(adj_normalized) def construct_feed_dict(features, support, labels, labels_mask, placeholders): """Construct feed dictionary.""" feed_dict = dict() feed_dict.update({placeholders['labels']: labels}) feed_dict.update({placeholders['labels_mask']: labels_mask}) feed_dict.update({placeholders['features']: features}) feed_dict.update({placeholders['support'][i]: support[i] for i in range(len(support))}) feed_dict.update({placeholders['num_features_nonzero']: features[1].shape}) return feed_dict def chebyshev_polynomials(adj, k): """Calculate Chebyshev polynomials up to order k. Return a list of sparse matrices (tuple representation).""" print("Calculating Chebyshev polynomials up to order {}...".format(k)) adj_normalized = normalize_adj(adj) laplacian = sp.eye(adj.shape[0]) - adj_normalized largest_eigval, _ = eigsh(laplacian, 1, which='LM') scaled_laplacian = (2. / largest_eigval[0]) * laplacian - sp.eye(adj.shape[0]) t_k = list() t_k.append(sp.eye(adj.shape[0])) t_k.append(scaled_laplacian) def chebyshev_recurrence(t_k_minus_one, t_k_minus_two, scaled_lap): s_lap = sp.csr_matrix(scaled_lap, copy=True) return 2 * s_lap.dot(t_k_minus_one) - t_k_minus_two for i in range(2, k+1): t_k.append(chebyshev_recurrence(t_k[-1], t_k[-2], scaled_laplacian)) return sparse_to_tuple(t_k)
6,500
37.241176
115
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/layers.py
from gcn.inits import * import tensorflow as tf flags = tf.app.flags FLAGS = flags.FLAGS # global unique layer ID dictionary for layer name assignment _LAYER_UIDS = {} def get_layer_uid(layer_name=''): """Helper function, assigns unique layer IDs.""" if layer_name not in _LAYER_UIDS: _LAYER_UIDS[layer_name] = 1 return 1 else: _LAYER_UIDS[layer_name] += 1 return _LAYER_UIDS[layer_name] def sparse_dropout(x, keep_prob, noise_shape): """Dropout for sparse tensors.""" random_tensor = keep_prob random_tensor += tf.random_uniform(noise_shape) dropout_mask = tf.cast(tf.floor(random_tensor), dtype=tf.bool) pre_out = tf.sparse_retain(x, dropout_mask) return pre_out * (1./keep_prob) def dot(x, y, sparse=False): """Wrapper for tf.matmul (sparse vs dense).""" if sparse: res = tf.sparse_tensor_dense_matmul(x, y) else: res = tf.matmul(x, y) return res class Layer(object): """Base layer class. Defines basic API for all layer objects. Implementation inspired by keras (http://keras.io). # Properties name: String, defines the variable scope of the layer. logging: Boolean, switches Tensorflow histogram logging on/off # Methods _call(inputs): Defines computation graph of layer (i.e. takes input, returns output) __call__(inputs): Wrapper for _call() _log_vars(): Log all variables """ def __init__(self, **kwargs): allowed_kwargs = {'name', 'logging'} for kwarg in kwargs.keys(): assert kwarg in allowed_kwargs, 'Invalid keyword argument: ' + kwarg name = kwargs.get('name') if not name: layer = self.__class__.__name__.lower() name = layer + '_' + str(get_layer_uid(layer)) self.name = name self.vars = {} logging = kwargs.get('logging', False) self.logging = logging self.sparse_inputs = False def _call(self, inputs): return inputs def __call__(self, inputs): with tf.name_scope(self.name): if self.logging and not self.sparse_inputs: tf.summary.histogram(self.name + '/inputs', inputs) outputs = self._call(inputs) if self.logging: tf.summary.histogram(self.name + '/outputs', outputs) return outputs def _log_vars(self): for var in self.vars: tf.summary.histogram(self.name + '/vars/' + var, self.vars[var]) class Dense(Layer): """Dense layer.""" def __init__(self, input_dim, output_dim, placeholders, dropout=0., sparse_inputs=False, act=tf.nn.relu, bias=False, featureless=False, **kwargs): super(Dense, self).__init__(**kwargs) if dropout: self.dropout = placeholders['dropout'] else: self.dropout = 0. self.act = act self.sparse_inputs = sparse_inputs self.featureless = featureless self.bias = bias # helper variable for sparse dropout self.num_features_nonzero = placeholders['num_features_nonzero'] with tf.variable_scope(self.name + '_vars'): self.vars['weights'] = glorot([input_dim, output_dim], name='weights') if self.bias: self.vars['bias'] = zeros([output_dim], name='bias') if self.logging: self._log_vars() def _call(self, inputs): x = inputs # dropout if self.sparse_inputs: x = sparse_dropout(x, 1-self.dropout, self.num_features_nonzero) else: x = tf.nn.dropout(x, 1-self.dropout) # transform output = dot(x, self.vars['weights'], sparse=self.sparse_inputs) # bias if self.bias: output += self.vars['bias'] return self.act(output) class GraphConvolution(Layer): """Graph convolution layer.""" def __init__(self, input_dim, output_dim, placeholders, dropout=0., sparse_inputs=False, act=tf.nn.relu, bias=False, featureless=False, **kwargs): super(GraphConvolution, self).__init__(**kwargs) if dropout: self.dropout = placeholders['dropout'] else: self.dropout = 0. self.act = act self.support = placeholders['support'] self.sparse_inputs = sparse_inputs self.featureless = featureless self.bias = bias # helper variable for sparse dropout self.num_features_nonzero = placeholders['num_features_nonzero'] with tf.variable_scope(self.name + '_vars'): for i in range(len(self.support)): self.vars['weights_' + str(i)] = glorot([input_dim, output_dim], name='weights_' + str(i)) if self.bias: self.vars['bias'] = zeros([output_dim], name='bias') if self.logging: self._log_vars() def _call(self, inputs): x = inputs # dropout if self.sparse_inputs: x = sparse_dropout(x, 1-self.dropout, self.num_features_nonzero) else: x = tf.nn.dropout(x, 1-self.dropout) # convolve supports = list() for i in range(len(self.support)): if not self.featureless: pre_sup = dot(x, self.vars['weights_' + str(i)], sparse=self.sparse_inputs) else: pre_sup = self.vars['weights_' + str(i)] support = dot(self.support[i], pre_sup, sparse=True) supports.append(support) output = tf.add_n(supports) # bias if self.bias: output += self.vars['bias'] return self.act(output)
5,886
30.148148
92
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/models.py
from gcn.layers import * from gcn.metrics import * flags = tf.app.flags FLAGS = flags.FLAGS class Model(object): def __init__(self, **kwargs): allowed_kwargs = {'name', 'logging'} for kwarg in kwargs.keys(): assert kwarg in allowed_kwargs, 'Invalid keyword argument: ' + kwarg name = kwargs.get('name') if not name: name = self.__class__.__name__.lower() self.name = name logging = kwargs.get('logging', False) self.logging = logging self.vars = {} self.placeholders = {} self.layers = [] self.activations = [] self.inputs = None self.outputs = None self.loss = 0 self.accuracy = 0 self.optimizer = None self.opt_op = None def _build(self): raise NotImplementedError def build(self): """ Wrapper for _build() """ with tf.variable_scope(self.name): self._build() # Build sequential layer model self.activations.append(self.inputs) for layer in self.layers: hidden = layer(self.activations[-1]) self.activations.append(hidden) self.outputs = self.activations[-1] # Store model variables for easy access variables = tf.get_collection(tf.GraphKeys.GLOBAL_VARIABLES, scope=self.name) self.vars = {var.name: var for var in variables} # Build metrics self._loss() self._accuracy() self.opt_op = self.optimizer.minimize(self.loss) def predict(self): pass def _loss(self): raise NotImplementedError def _accuracy(self): raise NotImplementedError def save(self, sess=None): if not sess: raise AttributeError("TensorFlow session not provided.") saver = tf.train.Saver(self.vars) save_path = saver.save(sess, "tmp/%s.ckpt" % self.name) print("Model saved in file: %s" % save_path) def load(self, sess=None): if not sess: raise AttributeError("TensorFlow session not provided.") saver = tf.train.Saver(self.vars) save_path = "tmp/%s.ckpt" % self.name saver.restore(sess, save_path) print("Model restored from file: %s" % save_path) class MLP(Model): def __init__(self, placeholders, input_dim, **kwargs): super(MLP, self).__init__(**kwargs) self.inputs = placeholders['features'] self.input_dim = input_dim # self.input_dim = self.inputs.get_shape().as_list()[1] # To be supported in future Tensorflow versions self.output_dim = placeholders['labels'].get_shape().as_list()[1] self.placeholders = placeholders self.optimizer = tf.train.AdamOptimizer(learning_rate=FLAGS.learning_rate) self.build() def _loss(self): # Weight decay loss for var in self.layers[0].vars.values(): self.loss += FLAGS.weight_decay * tf.nn.l2_loss(var) # Cross entropy error self.loss += masked_softmax_cross_entropy(self.outputs, self.placeholders['labels'], self.placeholders['labels_mask']) def _accuracy(self): self.accuracy = masked_accuracy(self.outputs, self.placeholders['labels'], self.placeholders['labels_mask']) def _build(self): self.layers.append(Dense(input_dim=self.input_dim, output_dim=FLAGS.hidden1, placeholders=self.placeholders, act=tf.nn.relu, dropout=True, sparse_inputs=True, logging=self.logging)) self.layers.append(Dense(input_dim=FLAGS.hidden1, output_dim=self.output_dim, placeholders=self.placeholders, act=lambda x: x, dropout=True, logging=self.logging)) def predict(self): return tf.nn.softmax(self.outputs) class GCN(Model): def __init__(self, placeholders, input_dim, **kwargs): super(GCN, self).__init__(**kwargs) self.inputs = placeholders['features'] self.input_dim = input_dim # self.input_dim = self.inputs.get_shape().as_list()[1] # To be supported in future Tensorflow versions self.output_dim = placeholders['labels'].get_shape().as_list()[1] self.placeholders = placeholders self.optimizer = tf.train.AdamOptimizer(learning_rate=FLAGS.learning_rate) self.build() def _loss(self): # Weight decay loss for var in self.layers[0].vars.values(): self.loss += FLAGS.weight_decay * tf.nn.l2_loss(var) # Cross entropy error self.loss += masked_softmax_cross_entropy(self.outputs, self.placeholders['labels'], self.placeholders['labels_mask']) def _accuracy(self): self.accuracy = masked_accuracy(self.outputs, self.placeholders['labels'], self.placeholders['labels_mask']) def _build(self): self.layers.append(GraphConvolution(input_dim=self.input_dim, output_dim=FLAGS.hidden1, placeholders=self.placeholders, act=tf.nn.relu, dropout=True, sparse_inputs=True, logging=self.logging)) self.layers.append(GraphConvolution(input_dim=FLAGS.hidden1, output_dim=self.output_dim, placeholders=self.placeholders, act=lambda x: x, dropout=True, logging=self.logging)) def predict(self): return tf.nn.softmax(self.outputs)
6,264
34.196629
112
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/metrics.py
import tensorflow as tf def masked_softmax_cross_entropy(preds, labels, mask): """Softmax cross-entropy loss with masking.""" loss = tf.nn.softmax_cross_entropy_with_logits(logits=preds, labels=labels) mask = tf.cast(mask, dtype=tf.float32) mask /= tf.reduce_mean(mask) loss *= mask return tf.reduce_mean(loss) def masked_accuracy(preds, labels, mask): """Accuracy with masking.""" correct_prediction = tf.equal(tf.argmax(preds, 1), tf.argmax(labels, 1)) accuracy_all = tf.cast(correct_prediction, tf.float32) mask = tf.cast(mask, dtype=tf.float32) mask /= tf.reduce_mean(mask) accuracy_all *= mask return tf.reduce_mean(accuracy_all)
691
31.952381
79
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/__init__.py
from __future__ import print_function from __future__ import division
70
22.666667
37
py
gcnn-survey-paper
gcnn-survey-paper-master/third_party/gcn/gcn/train.py
from __future__ import division from __future__ import print_function import time import tensorflow as tf from gcn.utils import * from gcn.models import GCN, MLP # Set random seed seed = 123 np.random.seed(seed) tf.set_random_seed(seed) # Settings flags = tf.app.flags FLAGS = flags.FLAGS flags.DEFINE_string('dataset', 'cora', 'Dataset string.') # 'cora', 'citeseer', 'pubmed' flags.DEFINE_string('model', 'gcn', 'Model string.') # 'gcn', 'gcn_cheby', 'dense' flags.DEFINE_float('learning_rate', 0.01, 'Initial learning rate.') flags.DEFINE_integer('epochs', 200, 'Number of epochs to train.') flags.DEFINE_integer('hidden1', 16, 'Number of units in hidden layer 1.') flags.DEFINE_float('dropout', 0.5, 'Dropout rate (1 - keep probability).') flags.DEFINE_float('weight_decay', 5e-4, 'Weight for L2 loss on embedding matrix.') flags.DEFINE_integer('early_stopping', 10, 'Tolerance for early stopping (# of epochs).') flags.DEFINE_integer('max_degree', 3, 'Maximum Chebyshev polynomial degree.') # Load data adj, features, y_train, y_val, y_test, train_mask, val_mask, test_mask = load_data(FLAGS.dataset) # Some preprocessing features = preprocess_features(features) if FLAGS.model == 'gcn': support = [preprocess_adj(adj)] num_supports = 1 model_func = GCN elif FLAGS.model == 'gcn_cheby': support = chebyshev_polynomials(adj, FLAGS.max_degree) num_supports = 1 + FLAGS.max_degree model_func = GCN elif FLAGS.model == 'dense': support = [preprocess_adj(adj)] # Not used num_supports = 1 model_func = MLP else: raise ValueError('Invalid argument for model: ' + str(FLAGS.model)) # Define placeholders placeholders = { 'support': [tf.sparse_placeholder(tf.float32) for _ in range(num_supports)], 'features': tf.sparse_placeholder(tf.float32, shape=tf.constant(features[2], dtype=tf.int64)), 'labels': tf.placeholder(tf.float32, shape=(None, y_train.shape[1])), 'labels_mask': tf.placeholder(tf.int32), 'dropout': tf.placeholder_with_default(0., shape=()), 'num_features_nonzero': tf.placeholder(tf.int32) # helper variable for sparse dropout } # Create model model = model_func(placeholders, input_dim=features[2][1], logging=True) # Initialize session sess = tf.Session() # Define model evaluation function def evaluate(features, support, labels, mask, placeholders): t_test = time.time() feed_dict_val = construct_feed_dict(features, support, labels, mask, placeholders) outs_val = sess.run([model.loss, model.accuracy], feed_dict=feed_dict_val) return outs_val[0], outs_val[1], (time.time() - t_test) # Init variables sess.run(tf.global_variables_initializer()) cost_val = [] # Train model for epoch in range(FLAGS.epochs): t = time.time() # Construct feed dictionary feed_dict = construct_feed_dict(features, support, y_train, train_mask, placeholders) feed_dict.update({placeholders['dropout']: FLAGS.dropout}) # Training step outs = sess.run([model.opt_op, model.loss, model.accuracy], feed_dict=feed_dict) # Validation cost, acc, duration = evaluate(features, support, y_val, val_mask, placeholders) cost_val.append(cost) # Print results print("Epoch:", '%04d' % (epoch + 1), "train_loss=", "{:.5f}".format(outs[1]), "train_acc=", "{:.5f}".format(outs[2]), "val_loss=", "{:.5f}".format(cost), "val_acc=", "{:.5f}".format(acc), "time=", "{:.5f}".format(time.time() - t)) if epoch > FLAGS.early_stopping and cost_val[-1] > np.mean(cost_val[-(FLAGS.early_stopping+1):-1]): print("Early stopping...") break print("Optimization Finished!") # Testing test_cost, test_acc, test_duration = evaluate(features, support, y_test, test_mask, placeholders) print("Test set results:", "cost=", "{:.5f}".format(test_cost), "accuracy=", "{:.5f}".format(test_acc), "time=", "{:.5f}".format(test_duration))
3,891
35.037037
103
py
gcnn-survey-paper
gcnn-survey-paper-master/models/node_edge_models.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Inference step for joint node classification and link prediction models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from models.base_models import NodeEdgeModel from models.edge_models import Gae from models.node_models import Gat from models.node_models import Gcn import tensorflow as tf from utils.model_utils import compute_adj from utils.model_utils import gat_module from utils.model_utils import gcn_module from utils.model_utils import get_sp_topk from utils.model_utils import mask_edges class GaeGat(NodeEdgeModel): """GAE for link prediction and GAT for node classification.""" def __init__(self, config): """Initializes EGCNGAT model.""" super(GaeGat, self).__init__(config) self.edge_model = Gae(config) self.node_model = Gat(config) def compute_inference(self, node_features_in, sp_adj_matrix, is_training): adj_matrix_pred = self.edge_model.compute_inference( node_features_in, sp_adj_matrix, is_training) self.adj_matrix_pred = adj_matrix_pred adj_mask = get_sp_topk(adj_matrix_pred, sp_adj_matrix, self.nb_nodes, self.topk) self.adj_mask = adj_mask # masked_adj_matrix_pred = tf.multiply(adj_mask, # tf.nn.sigmoid(adj_matrix_pred)) masked_adj_matrix_pred = mask_edges(tf.nn.sigmoid(adj_matrix_pred), adj_mask) sp_adj_pred = tf.contrib.layers.dense_to_sparse(masked_adj_matrix_pred) logits = self.node_model.compute_inference(node_features_in, sp_adj_pred, is_training) return logits, adj_matrix_pred class GaeGcn(NodeEdgeModel): """GAE for link prediction and GCN for node classification.""" def __init__(self, config): """Initializes EGCNGCN model.""" super(GaeGcn, self).__init__(config) self.edge_model = Gae(config) self.node_model = Gcn(config) def compute_inference(self, node_features_in, sp_adj_matrix, is_training): adj_matrix_pred = self.edge_model.compute_inference( node_features_in, sp_adj_matrix, is_training) self.adj_matrix_pred = adj_matrix_pred adj_mask = get_sp_topk(adj_matrix_pred, sp_adj_matrix, self.nb_nodes, self.topk) sp_adj_pred = tf.contrib.layers.dense_to_sparse( tf.multiply(adj_mask, tf.nn.leaky_relu(adj_matrix_pred))) sp_adj_pred = tf.sparse_softmax(sp_adj_pred) logits = self.node_model.compute_inference(node_features_in, sp_adj_pred, is_training) return logits, adj_matrix_pred ############################ EXPERIMENTAL MODELS ############################# class GatGraphite(NodeEdgeModel): """Gae for link prediction and GCN for node classification.""" def compute_inference(self, node_features_in, sp_adj_matrix, is_training): with tf.variable_scope('edge-model'): z_latent = gat_module( node_features_in, sp_adj_matrix, self.n_hidden_edge, self.n_att_edge, self.p_drop_edge, is_training, self.input_dim, self.sparse_features, average_last=False) adj_matrix_pred = compute_adj(z_latent, self.att_mechanism, self.p_drop_edge, is_training) self.adj_matrix_pred = adj_matrix_pred with tf.variable_scope('node-model'): concat = True if concat: z_latent = tf.sparse_concat( axis=1, sp_inputs=[ tf.contrib.layers.dense_to_sparse(z_latent), node_features_in ], ) sparse_features = True input_dim = self.n_hidden_edge[-1] * self.n_att_edge[ -1] + self.input_dim else: sparse_features = False input_dim = self.n_hidden_edge[-1] * self.n_att_edge[-1] logits = gat_module( z_latent, sp_adj_matrix, self.n_hidden_node, self.n_att_node, self.p_drop_node, is_training, input_dim, sparse_features=sparse_features, average_last=False) return logits, adj_matrix_pred class GaeGatConcat(NodeEdgeModel): """EGCN for link prediction and GCN for node classification.""" def __init__(self, config): """Initializes EGCN_GAT model.""" super(GaeGatConcat, self).__init__(config) self.edge_model = Gae(config) self.node_model = Gat(config) def compute_inference(self, node_features_in, sp_adj_matrix, is_training): with tf.variable_scope('edge-model'): z_latent = gcn_module(node_features_in, sp_adj_matrix, self.n_hidden_edge, self.p_drop_edge, is_training, self.input_dim, self.sparse_features) adj_matrix_pred = compute_adj(z_latent, self.att_mechanism, self.p_drop_edge, is_training) self.adj_matrix_pred = adj_matrix_pred with tf.variable_scope('node-model'): z_latent = tf.sparse_concat( axis=1, sp_inputs=[ tf.contrib.layers.dense_to_sparse(z_latent), node_features_in ]) sparse_features = True input_dim = self.n_hidden_edge[-1] + self.input_dim sp_adj_train = tf.SparseTensor( indices=sp_adj_matrix.indices, values=tf.ones_like(sp_adj_matrix.values), dense_shape=sp_adj_matrix.dense_shape) logits = gat_module( z_latent, sp_adj_train, self.n_hidden_node, self.n_att_node, self.p_drop_node, is_training, input_dim, sparse_features=sparse_features, average_last=True) return logits, adj_matrix_pred class GaeGcnConcat(NodeEdgeModel): """EGCN for link prediction and GCN for node classification.""" def compute_inference(self, node_features_in, sp_adj_matrix, is_training): with tf.variable_scope('edge-model'): z_latent = gcn_module(node_features_in, sp_adj_matrix, self.n_hidden_edge, self.p_drop_edge, is_training, self.input_dim, self.sparse_features) adj_matrix_pred = compute_adj(z_latent, self.att_mechanism, self.p_drop_edge, is_training) self.adj_matrix_pred = adj_matrix_pred with tf.variable_scope('node-model'): z_latent = tf.sparse_concat( axis=1, sp_inputs=[ tf.contrib.layers.dense_to_sparse(z_latent), node_features_in ]) sparse_features = True input_dim = self.n_hidden_edge[-1] + self.input_dim logits = gcn_module( z_latent, sp_adj_matrix, self.n_hidden_node, self.p_drop_node, is_training, input_dim, sparse_features=sparse_features) return logits, adj_matrix_pred class Gcat(NodeEdgeModel): """1 iteration Graph Convolution Attention Model.""" def __init__(self, config): """Initializes GCAT model.""" super(Gcat, self).__init__(config) self.edge_model = Gae(config) self.node_model = Gcn(config) def compute_inference(self, node_features_in, sp_adj_matrix, is_training): """Forward pass for GAT model.""" adj_matrix_pred = self.edge_model.compute_inference( node_features_in, sp_adj_matrix, is_training) sp_adj_mask = tf.SparseTensor( indices=sp_adj_matrix.indices, values=tf.ones_like(sp_adj_matrix.values), dense_shape=sp_adj_matrix.dense_shape) sp_adj_att = sp_adj_mask * adj_matrix_pred sp_adj_att = tf.SparseTensor( indices=sp_adj_att.indices, values=tf.nn.leaky_relu(sp_adj_att.values), dense_shape=sp_adj_att.dense_shape) sp_adj_att = tf.sparse_softmax(sp_adj_att) logits = self.node_model.compute_inference(node_features_in, sp_adj_att, is_training) return logits, adj_matrix_pred
8,619
36.155172
80
py
gcnn-survey-paper
gcnn-survey-paper-master/models/edge_models.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Inference step for link prediction models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from models.base_models import EdgeModel import tensorflow as tf from utils.model_utils import compute_adj from utils.model_utils import gat_module from utils.model_utils import gcn_module from utils.model_utils import mlp_module class Gae(EdgeModel): """Graph Auto-Encoder (GAE) (Kipf & al) for link prediction. arXiv link: https://arxiv.org/abs/1611.07308 """ def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for GAE model.""" sparse = self.sparse_features in_dim = self.input_dim with tf.variable_scope('edge-model'): h0 = gcn_module(node_features, adj_matrix, self.n_hidden, self.p_drop, is_training, in_dim, sparse) adj_matrix_pred = compute_adj(h0, self.att_mechanism, self.p_drop, is_training) self.adj_matrix_pred = tf.nn.sigmoid(adj_matrix_pred) return adj_matrix_pred class Egat(EdgeModel): """Edge-GAT for link prediction.""" def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for GAE model.""" sparse = self.sparse_features in_dim = self.input_dim with tf.variable_scope('edge-model'): h0 = gat_module( node_features, adj_matrix, self.n_hidden, self.n_att, self.p_drop, is_training, in_dim, sparse, average_last=True) adj_matrix_pred = compute_adj(h0, self.att_mechanism, self.p_drop, is_training) self.adj_matrix_pred = tf.nn.sigmoid(adj_matrix_pred) return adj_matrix_pred class Vgae(EdgeModel): """Variational Graph Auto-Encoder (VGAE) (Kipf & al) for link prediction. arXiv link: https://arxiv.org/abs/1611.07308 """ def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for GAE model.""" sparse = self.sparse_features in_dim = self.input_dim with tf.variable_scope('edge-model'): h0 = gcn_module(node_features, adj_matrix, self.n_hidden[:-1], self.p_drop, is_training, in_dim, sparse) # N x F with tf.variable_scope('mean'): z_mean = gcn_module(h0, adj_matrix, self.n_hidden[-1:], self.p_drop, is_training, self.n_hidden[-2], False) self.z_mean = z_mean with tf.variable_scope('std'): # N x F z_log_std = gcn_module(h0, adj_matrix, self.n_hidden[-1:], self.p_drop, is_training, self.n_hidden[-2], False) self.z_log_std = z_log_std # add noise during training noise = tf.random_normal([self.nb_nodes, self.n_hidden[-1] ]) * tf.exp(z_log_std) z = tf.cond(is_training, lambda: tf.add(z_mean, noise), lambda: z_mean) # N x N adj_matrix_pred = compute_adj(z, self.att_mechanism, self.p_drop, is_training) self.adj_matrix_pred = tf.nn.sigmoid(adj_matrix_pred) return adj_matrix_pred def _compute_edge_loss(self, adj_pred, adj_train): """Overrides _compute_edge_loss to add Variational Inference objective.""" log_lik = super(Vgae, self)._compute_edge_loss(adj_pred, adj_train) norm = self.nb_nodes**2 / float((self.nb_nodes**2 - self.nb_edges) * 2) kl_mat = 0.5 * tf.reduce_sum( 1 + 2 * self.z_log_std - tf.square(self.z_mean) - tf.square( tf.exp(self.z_log_std)), 1) kl = tf.reduce_mean(kl_mat) / self.nb_nodes edge_loss = norm * log_lik - kl return edge_loss class Emlp(EdgeModel): """Simple baseline for link prediction. Creates a tensorflow graph to train and evaluate EMLP on graph data. """ def compute_inference(self, node_features, _, is_training): """Forward step for GAE model.""" sparse = self.sparse_features in_dim = self.input_dim with tf.variable_scope('edge-model'): h0 = mlp_module( node_features, self.n_hidden, self.p_drop, is_training, in_dim, sparse, use_bias=False) adj_matrix_pred = compute_adj(h0, self.att_mechanism, self.p_drop, is_training) self.adj_matrix_pred = tf.nn.sigmoid(adj_matrix_pred) return adj_matrix_pred
5,060
34.640845
79
py
gcnn-survey-paper
gcnn-survey-paper-master/models/node_models.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Inference step for node classification models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from models.base_models import NodeModel import tensorflow as tf from utils.model_utils import cheby_module from utils.model_utils import compute_adj from utils.model_utils import gat_module from utils.model_utils import gcn_module from utils.model_utils import gcn_pool_layer from utils.model_utils import mlp_module from utils.model_utils import sp_gat_layer from utils.model_utils import sp_gcn_layer class Gat(NodeModel): """Graph Attention (GAT) Model (Velickovic & al). arXiv link: https://arxiv.org/abs/1710.10903 """ def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for GAT model.""" sparse = self.sparse_features in_dim = self.input_dim average_last = True with tf.variable_scope('node-model'): logits = gat_module(node_features, adj_matrix, self.n_hidden, self.n_att, self.p_drop, is_training, in_dim, sparse, average_last) return logits class Gcn(NodeModel): """Graph convolution network (Kipf & al). arXiv link: https://arxiv.org/abs/1609.02907 """ def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for graph convolution model.""" with tf.variable_scope('node-model'): logits = gcn_module(node_features, adj_matrix, self.n_hidden, self.p_drop, is_training, self.input_dim, self.sparse_features) return logits class Mlp(NodeModel): """Multi-layer perceptron model.""" def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for graph convolution model.""" with tf.variable_scope('node-model'): logits = mlp_module(node_features, self.n_hidden, self.p_drop, is_training, self.input_dim, self.sparse_features, use_bias=True) return logits class SemiEmb(NodeModel): """Deep Learning via Semi-Supervised Embedding (Weston & al). paper: http://icml2008.cs.helsinki.fi/papers/340.pdf """ def __init__(self, config): super(SemiEmb, self).__init__(config) self.semi_emb_k = config.semi_emb_k def compute_inference(self, node_features, adj_matrix, is_training): with tf.variable_scope('node-model'): hidden_repr = mlp_module(node_features, self.n_hidden, self.p_drop, is_training, self.input_dim, self.sparse_features, use_bias=True, return_hidden=True) logits = hidden_repr[-1] hidden_repr_reg = hidden_repr[self.semi_emb_k] l2_scores = compute_adj(hidden_repr_reg, self.att_mechanism, self.p_drop, is_training=False) self.l2_scores = tf.gather_nd(l2_scores, adj_matrix.indices) return logits def _compute_node_loss(self, logits, labels): supervised_loss = super(SemiEmb, self)._compute_node_loss(logits, labels) # supervised_loss = tf.nn.softmax_cross_entropy_with_logits( # labels=labels, logits=logits) # supervised_loss = tf.reduce_sum(supervised_loss) / self.nb_nodes reg_loss = tf.reduce_mean(self.l2_scores) return supervised_loss + self.edge_reg * reg_loss class Cheby(NodeModel): """Chebyshev polynomials for Spectral Graph Convolutions (Defferrard & al). arXiv link: https://arxiv.org/abs/1606.09375 """ def __init__(self, config): super(Cheby, self).__init__(config) self.cheby_k_loc = config.cheby_k_loc def compute_inference(self, node_features, normalized_laplacian, is_training): with tf.variable_scope('node-model'): dense_normalized_laplacian = tf.sparse_to_dense( sparse_indices=normalized_laplacian.indices, output_shape=normalized_laplacian.dense_shape, sparse_values=normalized_laplacian.values) cheby_polynomials = [tf.eye(self.nb_nodes), dense_normalized_laplacian] self.cheby = cheby_polynomials for _ in range(2, self.cheby_k_loc+1): cheby_polynomials.append(2 * tf.sparse_tensor_dense_matmul( normalized_laplacian, cheby_polynomials[-1]) - cheby_polynomials[-2] ) logits = cheby_module(node_features, cheby_polynomials, self.n_hidden, self.p_drop, is_training, self.input_dim, self.sparse_features) return logits ############################ EXPERIMENTAL MODELS ############################# class Hgat(NodeModel): """Hierarchical Graph Attention (GAT) Model.""" def compute_inference(self, node_features, adj_matrix, is_training): """Forward step for GAT model.""" in_dim = self.input_dim att = [] for j in range(4): with tf.variable_scope('gat-layer1-att{}'.format(j)): att.append( sp_gat_layer(node_features, adj_matrix, in_dim, 8, self.p_drop, is_training, True)) hidden_2 = [] hidden_2.append(tf.nn.elu(tf.concat(att[:2], axis=-1))) hidden_2.append(tf.nn.elu(tf.concat(att[2:], axis=-1))) att = [] for j in range(2): with tf.variable_scope('gat-layer2-att{}'.format(j)): att.append( sp_gat_layer(hidden_2[j], adj_matrix, 16, 7, self.p_drop, is_training, False)) return tf.add_n(att) / 2. class Pgcn(NodeModel): """Pooling Graph Convolution Network.""" def compute_inference(self, node_features, adj_matrix, is_training): adj_matrix_dense = tf.sparse_to_dense( sparse_indices=adj_matrix.indices, output_shape=adj_matrix.dense_shape, sparse_values=adj_matrix.values, validate_indices=False) adj_matrix_dense = tf.cast(tf.greater(adj_matrix_dense, 0), tf.float32) adj_matrix_dense = tf.expand_dims(adj_matrix_dense, -1) # N x N x 1 in_dim = self.input_dim sparse = self.sparse_features for i, out_dim in enumerate(self.n_hidden[:-1]): if i > 0: sparse = False with tf.variable_scope('gcn-pool-{}'.format(i)): node_features = gcn_pool_layer( node_features, adj_matrix_dense, in_dim=in_dim, out_dim=out_dim, sparse=sparse, is_training=is_training, p_drop=self.p_drop) node_features = tf.reshape(node_features, (-1, out_dim)) node_features = tf.contrib.layers.bias_add(node_features) node_features = tf.nn.elu(node_features) in_dim = out_dim with tf.variable_scope('gcn-layer-last'): logits = sp_gcn_layer(node_features, adj_matrix, in_dim, self.n_hidden[-1], self.p_drop, is_training, False) return logits
7,409
36.236181
80
py
gcnn-survey-paper
gcnn-survey-paper-master/models/__init__.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License.
565
39.428571
73
py
gcnn-survey-paper
gcnn-survey-paper-master/models/base_models.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Base models class. Main functionnalities for node classification models, link prediction models and joint node classification and link prediction models. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf class BaseModel(object): """Base model class. Defines basic functionnalities for all models.""" def __init__(self, config): """Initialize base model. Args: config: object of Config class defined in train.py, stores configuration parameters to build and train the model """ self.input_dim = config.input_dim self.lr = config.lr self.edge_reg = config.edge_reg self.edge_l2_reg = config.edge_l2_reg self.node_l2_reg = config.node_l2_reg self.nb_nodes = config.nb_nodes self.nb_edges = config.nb_edges self.sparse_features = config.sparse_features self.edge_loss = config.edge_loss self.att_mechanism = config.att_mechanism self.multilabel = config.multilabel def _create_placeholders(self): raise NotImplementedError def compute_inference(self, features, adj_matrix, is_training): raise NotImplementedError def build_graph(self): raise NotImplementedError def _create_optimizer(self, loss): """Create train operation.""" opt = tf.train.AdamOptimizer(learning_rate=self.lr) train_op = opt.minimize(loss) return train_op def _compute_node_loss(self, logits, labels): """Node classification loss with sigmoid cross entropy.""" if self.multilabel: loss = tf.nn.sigmoid_cross_entropy_with_logits( labels=labels, logits=logits) else: loss = tf.nn.softmax_cross_entropy_with_logits( labels=labels, logits=logits) return tf.reduce_mean(loss) def _compute_node_l2_loss(self): """L2 regularization loss for parameters in node classification model.""" all_variables = tf.trainable_variables() non_reg = ['bias', 'embeddings', 'beta', 'edge-model'] node_l2_loss = tf.add_n([ tf.nn.l2_loss(v) for v in all_variables if all([var_name not in v.name for var_name in non_reg]) ]) return node_l2_loss def _compute_edge_l2_loss(self): """L2 regularization loss for parameters in link prediction model.""" all_variables = tf.trainable_variables() edge_l2_loss = tf.add_n([tf.nn.l2_loss(v) for v in all_variables if \ 'edge-model' in v.name]) return edge_l2_loss def _compute_edge_loss_neg_sampling(self, adj_pred, adj_true): """Link prediction CE loss with negative sampling.""" keep_prob = self.nb_edges / (self.nb_nodes**2 - self.nb_edges) loss_mask = tf.nn.dropout( 1 - adj_true, keep_prob=keep_prob) * keep_prob loss_mask += adj_true boolean_mask = tf.greater(loss_mask, 0.) masked_pred = tf.boolean_mask(adj_pred, boolean_mask) masked_true = tf.boolean_mask(adj_true, boolean_mask) edge_loss = tf.nn.sigmoid_cross_entropy_with_logits( labels=masked_true, logits=masked_pred, ) return tf.reduce_mean(edge_loss) def _compute_edge_loss_weighted_ce(self, adj_pred, adj_true): """Link prediction loss with weighted sigmoid cross entropy.""" pos_weight = float((self.nb_nodes**2) - self.nb_edges) / self.nb_edges edge_loss = tf.nn.weighted_cross_entropy_with_logits( targets=adj_true, logits=adj_pred, pos_weight=pos_weight) return tf.reduce_mean(edge_loss) def _compute_edge_loss(self, adj_pred, adj_true): if self.edge_loss == 'weighted': return self._compute_edge_loss_weighted_ce(adj_pred, adj_true) else: return self._compute_edge_loss_neg_sampling(adj_pred, adj_true) class NodeModel(BaseModel): """Base model class for semi-supevised node classification.""" def __init__(self, config): """Initializes NodeModel for semi-supervised node classification. Args: config: object of Config class defined in train.py, stores configuration parameters to build and train the model """ super(NodeModel, self).__init__(config) self.p_drop = config.p_drop_node self.n_att = config.n_att_node self.n_hidden = config.n_hidden_node def _create_placeholders(self): """Create placeholders.""" with tf.name_scope('input'): self.placeholders = { 'adj_train': tf.sparse_placeholder(tf.float32), # normalized 'node_labels': tf.placeholder(tf.float32, shape=[None, self.n_hidden[-1]]), 'node_mask': tf.placeholder(tf.float32, shape=[ None, ]), 'is_training': tf.placeholder(tf.bool), } if self.sparse_features: self.placeholders['features'] = tf.sparse_placeholder(tf.float32) else: self.placeholders['features'] = tf.placeholder( tf.float32, shape=[None, self.input_dim]) def make_feed_dict(self, data, split, is_training): """Build feed dictionnary to train the model.""" feed_dict = { self.placeholders['adj_train']: data['adj_train_norm'], self.placeholders['features']: data['features'], self.placeholders['node_labels']: data['node_labels'], self.placeholders['node_mask']: data[split]['node_mask'], self.placeholders['is_training']: is_training } return feed_dict def build_graph(self): """Build tensorflow graph and create training, testing ops.""" self._create_placeholders() logits = self.compute_inference(self.placeholders['features'], self.placeholders['adj_train'], self.placeholders['is_training']) boolean_mask = tf.greater(self.placeholders['node_mask'], 0.) masked_pred = tf.boolean_mask(logits, boolean_mask) masked_true = tf.boolean_mask(self.placeholders['node_labels'], boolean_mask) loss = self._compute_node_loss(masked_pred, masked_true) loss += self.node_l2_reg * self._compute_node_l2_loss() train_op = self._create_optimizer(loss) metric_op, metric_update_op = self._create_metrics( masked_pred, masked_true) return loss, train_op, metric_op, metric_update_op def _create_metrics(self, logits, node_labels): """Create evaluation metrics for node classification.""" with tf.name_scope('metrics'): metrics = {} if self.multilabel: predictions = tf.cast( tf.greater(tf.nn.sigmoid(logits), 0.5), tf.float32) metrics['recall'], rec_op = tf.metrics.recall( labels=node_labels, predictions=predictions) metrics['precision'], prec_op = tf.metrics.precision( labels=node_labels, predictions=predictions) metrics['f1'] = 2 * metrics['precision'] * metrics['recall'] / ( metrics['precision'] + metrics['recall'] ) update_ops = [rec_op, prec_op] else: metrics['node_acc'], acc_op = tf.metrics.accuracy( labels=tf.argmax(node_labels, 1), predictions=tf.argmax(logits, 1)) update_ops = [acc_op] return metrics, update_ops class EdgeModel(BaseModel): """Base model class for link prediction.""" def __init__(self, config): """Initializes Edge model for link prediction. Args: config: object of Config class defined in train.py, stores configuration parameters to build and train the model """ super(EdgeModel, self).__init__(config) self.p_drop = config.p_drop_edge self.n_att = config.n_att_edge self.n_hidden = config.n_hidden_edge def _create_placeholders(self): """Create placeholders.""" with tf.name_scope('input'): self.placeholders = { # to compute metrics 'adj_true': tf.placeholder(tf.float32, shape=[None, None]), # to compute loss 'adj_train': tf.placeholder(tf.float32, shape=[None, None]), # for inference step 'adj_train_norm': tf.sparse_placeholder(tf.float32), # normalized 'edge_mask': tf.sparse_placeholder(tf.float32), 'is_training': tf.placeholder(tf.bool), } if self.sparse_features: self.placeholders['features'] = tf.sparse_placeholder(tf.float32) else: self.placeholders['features'] = tf.placeholder( tf.float32, shape=[None, self.input_dim]) def make_feed_dict(self, data, split, is_training): """Build feed dictionnary to train the model.""" feed_dict = { self.placeholders['features']: data['features'], self.placeholders['adj_true']: data['adj_true'], self.placeholders['adj_train']: data['adj_train'], self.placeholders['adj_train_norm']: data['adj_train_norm'], self.placeholders['edge_mask']: data[split]['edge_mask'], self.placeholders['is_training']: is_training } return feed_dict def build_graph(self): """Build tensorflow graph and create training, testing ops.""" self._create_placeholders() adj_pred = self.compute_inference(self.placeholders['features'], self.placeholders['adj_train_norm'], self.placeholders['is_training']) adj_train = tf.reshape(self.placeholders['adj_train'], (-1,)) loss = self._compute_edge_loss(tf.reshape(adj_pred, (-1,)), adj_train) loss += self.edge_l2_reg * self._compute_edge_l2_loss() train_op = self._create_optimizer(loss) masked_true = tf.reshape(tf.gather_nd( self.placeholders['adj_true'], self.placeholders['edge_mask'].indices), (-1,)) masked_pred = tf.reshape(tf.gather_nd( adj_pred, self.placeholders['edge_mask'].indices), (-1,)) metric_op, metric_update_op = self._create_metrics(masked_pred, masked_true) return loss, train_op, metric_op, metric_update_op def _create_metrics(self, adj_pred, adj_true): """Create evaluation metrics for node classification.""" with tf.name_scope('metrics'): metrics = {} metrics['edge_roc_auc'], roc_op = tf.metrics.auc( labels=adj_true, predictions=tf.sigmoid(adj_pred), curve='ROC' ) metrics['edge_pr_auc'], pr_op = tf.metrics.auc( labels=adj_true, predictions=tf.sigmoid(adj_pred), curve='PR' ) update_ops = [roc_op, pr_op] return metrics, update_ops class NodeEdgeModel(BaseModel): """Model class for semi-supevised node classification and link prediction.""" def __init__(self, config): """Initializes model. Args: config: object of Config class defined in train.py, stores configuration parameters to build and train the model """ super(NodeEdgeModel, self).__init__(config) self.n_att_edge = config.n_att_edge self.n_hidden_edge = config.n_hidden_edge self.p_drop_edge = config.p_drop_edge self.n_att_node = config.n_att_node self.n_hidden_node = config.n_hidden_node self.p_drop_node = config.p_drop_node self.topk = config.topk def _create_placeholders(self): """Create placeholders.""" with tf.name_scope('input'): self.placeholders = { 'adj_true': tf.placeholder(tf.float32, shape=[None, None]), # to compute loss 'adj_train': tf.placeholder(tf.float32, shape=[None, None]), # for inference step 'adj_train_norm': tf.sparse_placeholder(tf.float32), # normalized 'edge_mask': tf.sparse_placeholder(tf.float32), 'node_labels': tf.placeholder(tf.float32, shape=[None, self.n_hidden_node[-1]]), 'node_mask': tf.placeholder(tf.float32, shape=[ None, ]), 'is_training': tf.placeholder(tf.bool), } if self.sparse_features: self.placeholders['features'] = tf.sparse_placeholder(tf.float32) else: self.placeholders['features'] = tf.placeholder( tf.float32, shape=[None, self.input_dim]) def make_feed_dict(self, data, split, is_training): """Build feed dictionnary to train the model.""" feed_dict = { self.placeholders['features']: data['features'], self.placeholders['adj_true']: data['adj_true'], self.placeholders['adj_train']: data['adj_train'], self.placeholders['adj_train_norm']: data['adj_train_norm'], self.placeholders['edge_mask']: data[split]['edge_mask'], self.placeholders['node_labels']: data['node_labels'], self.placeholders['node_mask']: data[split]['node_mask'], self.placeholders['is_training']: is_training } return feed_dict def build_graph(self): """Build tensorflow graph and create training, testing ops.""" self._create_placeholders() logits, adj_pred = self.compute_inference( self.placeholders['features'], self.placeholders['adj_train_norm'], self.placeholders['is_training']) adj_train = tf.reshape(self.placeholders['adj_train'], (-1,)) boolean_node_mask = tf.greater(self.placeholders['node_mask'], 0.) masked_node_pred = tf.boolean_mask(logits, boolean_node_mask) masked_node_true = tf.boolean_mask(self.placeholders['node_labels'], boolean_node_mask) loss = self._compute_node_loss(masked_node_pred, masked_node_true) loss += self.node_l2_reg * self._compute_node_l2_loss() loss += self.edge_reg * self._compute_edge_loss( tf.reshape(adj_pred, (-1,)), adj_train) loss += self.edge_l2_reg * self._compute_edge_l2_loss() self.grad = tf.gradients(loss, self.adj_matrix_pred) train_op = self._create_optimizer(loss) masked_adj_true = tf.reshape(tf.gather_nd( self.placeholders['adj_true'], self.placeholders['edge_mask'].indices), (-1,)) masked_adj_pred = tf.reshape(tf.gather_nd( adj_pred, self.placeholders['edge_mask'].indices), (-1,)) metric_op, metric_update_op = self._create_metrics( masked_adj_pred, masked_adj_true, masked_node_pred, masked_node_true) return loss, train_op, metric_op, metric_update_op def _create_metrics(self, adj_pred, adj_true, node_pred, node_labels): """Create evaluation metrics for node classification.""" with tf.name_scope('metrics'): metrics = {} metrics['edge_roc_auc'], roc_op = tf.metrics.auc( labels=adj_true, predictions=tf.sigmoid(adj_pred), curve='ROC' ) metrics['edge_pr_auc'], pr_op = tf.metrics.auc( labels=adj_true, predictions=tf.sigmoid(adj_pred), curve='PR' ) metrics['node_acc'], acc_op = tf.metrics.accuracy( labels=tf.argmax(node_labels, 1), predictions=tf.argmax(node_pred, 1)) update_ops = [roc_op, pr_op, acc_op] return metrics, update_ops
15,593
38.180905
80
py
gcnn-survey-paper
gcnn-survey-paper-master/utils/model_utils.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Utils functions for GNN models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf WEIGHT_INIT = tf.contrib.layers.xavier_initializer() BIAS_INIT = tf.zeros_initializer() ############################## LAYERS ############################# def sparse_dropout(tensor, p_drop, is_training): """Dropout with sparse tensor.""" return tf.SparseTensor( indices=tensor.indices, values=tf.layers.dropout( inputs=tensor.values, rate=p_drop, training=is_training), dense_shape=tensor.dense_shape) def dense(node_features, in_dim, out_dim, p_drop, is_training, sparse, use_bias=False): """Dense layer with sparse or dense tensor and dropout.""" w_dense = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='linear', shape=(in_dim, out_dim)) if sparse: node_features = sparse_dropout(node_features, p_drop, is_training) node_features = tf.sparse_tensor_dense_matmul(node_features, w_dense) else: node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) node_features = tf.matmul(node_features, w_dense) if use_bias: node_features = tf.contrib.layers.bias_add(node_features) return node_features def sp_gcn_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse): """Single graph convolution layer with sparse tensors AXW. Args: node_features: Tensor of shape (nb_nodes, in_dim) or SparseTensor. adj_matrix: Sparse Tensor, normalized adjacency matrix. in_dim: integer specifying the input feature dimension. out_dim: integer specifying the output feature dimension. p_drop: dropout probability. is_training: boolean, True if the model is being trained, False otherwise. sparse: True if node_features are sparse. Returns: node_features: tensor of shape (nb_nodes, out_dim). New node features obtained from applying one GCN layer. Raises: """ node_features = dense(node_features, in_dim, out_dim, p_drop, is_training, sparse) node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) node_features = tf.sparse_tensor_dense_matmul(adj_matrix, node_features) return node_features def gcn_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse): """Single graph convolution layer with dense A. Args: node_features: Tensor of shape (nb_nodes, in_dim) or SparseTensor. adj_matrix: Tensor, normalized adjacency matrix. in_dim: integer specifying the input feature dimension. out_dim: integer specifying the output feature dimension. p_drop: dropout probability. is_training: boolean, True if the model is being trained, False otherwise. sparse: True if node_features are sparse. Returns: node_features: tensor of shape (nb_nodes, out_dim). New node features obtained from applying one GCN layer. Raises: """ node_features = dense(node_features, in_dim, out_dim, p_drop, is_training, sparse) node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) node_features = tf.matmul(adj_matrix, node_features) return node_features def gcn_pool_layer(node_features, adj_matrix, in_dim, out_dim, sparse, is_training, p_drop): """GCN with maxpooling over neighbours instead of avreaging.""" node_features = dense(node_features, in_dim, out_dim, p_drop, is_training, sparse) node_features = tf.expand_dims(node_features, 0) # 1 x N x d # broadcasting (adj in N x N x 1 and features are 1 x N x d) node_features = tf.multiply(node_features, adj_matrix) node_features = tf.transpose(node_features, perm=[0, 2, 1]) node_features = tf.reduce_max(node_features, axis=-1) # N x d return node_features def sp_gat_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse): """Single graph attention layer using sparse tensors. Args: node_features: Sparse Tensor of shape (nb_nodes, in_dim) or SparseTensor. adj_matrix: Sparse Tensor. in_dim: integer specifying the input feature dimension. out_dim: integer specifying the output feature dimension. p_drop: dropout probability. is_training: boolean, True if the model is being trained, False otherwise sparse: True if node features are sparse. Returns: node_features: tensor of shape (nb_nodes, out_dim). New node features obtained from applying one head of attention to input. Raises: """ # Linear transform node_features = dense(node_features, in_dim, out_dim, p_drop, is_training, sparse) # Attention scores alpha = sp_compute_adj_att(node_features, adj_matrix) alpha = tf.SparseTensor( indices=alpha.indices, values=tf.nn.leaky_relu(alpha.values), dense_shape=alpha.dense_shape) alpha = tf.sparse_softmax(alpha) alpha = sparse_dropout(alpha, p_drop, is_training) node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) # Compute self-attention features node_features = tf.sparse_tensor_dense_matmul(alpha, node_features) node_features = tf.contrib.layers.bias_add(node_features) return node_features def gat_layer(node_features, adj_matrix, out_dim, p_drop, is_training, i, j): """Single graph attention layer. Args: node_features: Tensor of shape (nb_nodes, feature_dim) adj_matrix: adjacency matrix. Tensor of shape (nb_nodes, nb_nodes) and type float. There should be 1 if there is a connection between two nodes and 0 otherwise. out_dim: integer specifying the output feature dimension. p_drop: dropout probability. is_training: boolean, True if the model is being trained, False otherwise i: layer index, used for naming variables j: attention mechanism index, used for naming variables Returns: node_features: tensor of shape (nb_nodes, out_dim). New node features obtained from applying one head of attention to input. Raises: """ with tf.variable_scope('gat-{}-{}'.format(i, j)): node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) # Linear transform of the features w_dense = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='linear', shape=(node_features.shape[1], out_dim)) node_features = tf.matmul(node_features, w_dense) alpha = compute_adj_att(node_features) alpha = tf.nn.leaky_relu(alpha) # Mask values before activation to inject the graph structure # Add -infinity to corresponding pairs before normalization bias_mat = -1e9 * (1. - adj_matrix) # multiply here if adjacency is weighted alpha = tf.nn.softmax(alpha + bias_mat, axis=-1) # alpha = tf.nn.softmax(alpha, axis=-1) alpha = tf.layers.dropout(inputs=alpha, rate=p_drop, training=is_training) node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) # Compute self-attention features node_features = tf.matmul(alpha, node_features) node_features = tf.contrib.layers.bias_add(node_features) return node_features def sp_egat_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse): """Single graph attention layer using sparse tensors. Args: node_features: Tensor of shape (nb_nodes, in_dim) or SparseTensor. adj_matrix: Sparse Tensor. in_dim: integer specifying the input feature dimension. out_dim: integer specifying the output feature dimension. p_drop: dropout probability. is_training: boolean, True if the model is being trained, False otherwise sparse: True if node features are sparse. Returns: node_features: tensor of shape (nb_nodes, out_dim). New node features obtained from applying one head of attention to input. Raises: """ # Linear transform node_features = dense(node_features, in_dim, out_dim, p_drop, is_training, sparse) # Attention scores alpha = sp_compute_adj_att(node_features, adj_matrix) alpha = tf.SparseTensor( indices=alpha.indices, values=tf.nn.leaky_relu(alpha.values), dense_shape=alpha.dense_shape) alpha = tf.sparse_softmax(alpha) alpha = sparse_dropout(alpha, p_drop, is_training) node_features = tf.layers.dropout( inputs=node_features, rate=p_drop, training=is_training) # Compute self-attention features node_features = tf.sparse_tensor_dense_matmul(alpha, node_features) node_features = tf.contrib.layers.bias_add(node_features) return node_features ############################## MULTI LAYERS ############################# def mlp_module(node_features, n_hidden, p_drop, is_training, in_dim, sparse_features, use_bias, return_hidden=False): """MLP.""" nb_layers = len(n_hidden) hidden_layers = [node_features] for i, out_dim in enumerate(n_hidden): with tf.variable_scope('mlp-{}'.format(i)): if i > 0: sparse_features = False if i == nb_layers - 1: use_bias = False h_i = dense(hidden_layers[-1], in_dim, out_dim, p_drop, is_training, sparse_features, use_bias) if i < nb_layers - 1: h_i = tf.nn.relu(h_i) in_dim = out_dim hidden_layers.append(h_i) if return_hidden: return hidden_layers else: return hidden_layers[-1] def gcn_module(node_features, adj_matrix, n_hidden, p_drop, is_training, in_dim, sparse_features): """GCN module with multiple layers.""" nb_layers = len(n_hidden) for i, out_dim in enumerate(n_hidden): if i > 0: sparse_features = False with tf.variable_scope('gcn-{}'.format(i)): node_features = sp_gcn_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse_features) if i < nb_layers - 1: node_features = tf.nn.relu(node_features) in_dim = out_dim return node_features def cheby_module(node_features, cheby_poly, n_hidden, p_drop, is_training, in_dim, sparse_features): """GCN module with multiple layers.""" nb_layers = len(n_hidden) for i, out_dim in enumerate(n_hidden): if i > 0: sparse_features = False feats = [] for j, poly in enumerate(cheby_poly): with tf.variable_scope('cheb-{}-{}'.format(i, j)): sparse_poly = tf.contrib.layers.dense_to_sparse(poly) feats.append(sp_gcn_layer(node_features, sparse_poly, in_dim, out_dim, p_drop, is_training, sparse_features)) node_features = tf.add_n(feats) if i < nb_layers - 1: node_features = tf.nn.relu(node_features) in_dim = out_dim return node_features def gat_module(node_features, adj_matrix, n_hidden, n_att, p_drop, is_training, in_dim, sparse_features, average_last): """GAT module with muli-headed attention and multiple layers.""" nb_layers = len(n_att) for i, k in enumerate(n_att): out_dim = n_hidden[i] att = [] if i > 0: sparse_features = False for j in range(k): with tf.variable_scope('gat-layer{}-att{}'.format(i, j)): att.append( sp_gat_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse_features)) # intermediate layers, concatenate features if i < nb_layers - 1: in_dim = out_dim * k node_features = tf.nn.elu(tf.concat(att, axis=-1)) if average_last: # last layer, average features instead of concatenating logits = tf.add_n(att) / n_att[-1] else: logits = tf.concat(att, axis=-1) return logits def egat_module(node_features, adj_matrix, n_hidden, n_att, p_drop, is_training, in_dim, sparse_features, average_last): """Edge-GAT module with muli-headed attention and multiple layers.""" nb_layers = len(n_att) for i, k in enumerate(n_att): out_dim = n_hidden[i] att = [] if i > 0: sparse_features = False for j in range(k): with tf.variable_scope('egat-layer{}-att{}'.format(i, j)): att.append( sp_gat_layer(node_features, adj_matrix, in_dim, out_dim, p_drop, is_training, sparse_features)) # intermediate layers, concatenate features if i < nb_layers - 1: in_dim = out_dim * k node_features = tf.nn.elu(tf.concat(att, axis=-1)) if average_last: # last layer, average features instead of concatenating logits = tf.add_n(att) / n_att[-1] else: logits = tf.concat(att, axis=-1) return logits ###################### EDGE SCORES FUNCTIONS ############################# def sp_compute_adj_att(node_features, adj_matrix_sp): """Self-attention for edges as in GAT with sparse adjacency.""" out_dim = node_features.shape[-1] # Self-attention mechanism a_row = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='selfatt-row', shape=(out_dim, 1)) a_col = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='selfatt-col', shape=(out_dim, 1)) alpha_row = tf.matmul(node_features, a_row) alpha_col = tf.matmul(node_features, a_col) # Compute matrix with self-attention scores using broadcasting alpha = tf.sparse_add(adj_matrix_sp * alpha_row, adj_matrix_sp * tf.transpose(alpha_col, perm=[1, 0])) return alpha def compute_adj_att(node_features): """Self-attention for edges as in GAT.""" out_dim = node_features.shape[-1] # Self-attention mechanism a_row = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='selfatt-row', shape=(out_dim, 1)) a_col = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='selfatt-col', shape=(out_dim, 1)) alpha_row = tf.matmul(node_features, a_row) alpha_col = tf.matmul(node_features, a_col) # Compute matrix with self-attention scores using broadcasting alpha = alpha_row + tf.transpose(alpha_col, perm=[1, 0]) # alpha += alpha_col + tf.transpose(alpha_row, perm=[1, 0]) return alpha def compute_weighted_mat_dot(node_features, nb_dots=1): """Compute weighted dot with matrix multiplication.""" adj_scores = [] in_dim = node_features.shape[-1] for i in range(nb_dots): weight_mat = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='w-dot-{}'.format(i), shape=(in_dim, in_dim)) adj_scores.append(tf.matmul(node_features, tf.matmul( weight_mat, tf.transpose(node_features, perm=[1, 0])))) return tf.add_n(adj_scores) def compute_weighted_dot(node_features, nb_dots=4): """Compute weighted dot product.""" adj_scores = [] in_dim = node_features.shape[-1] for i in range(nb_dots): weight_vec = tf.get_variable( initializer=WEIGHT_INIT, dtype=tf.float32, name='w-dot-{}'.format(i), shape=(1, in_dim)) weight_vec = tf.nn.softmax(weight_vec, axis=-1) adj_scores.append(tf.matmul(tf.multiply(weight_vec, node_features), tf.transpose(node_features, perm=[1, 0]))) return tf.add_n(adj_scores) def compute_l2_sim_matrix(node_features): """Compute squared-L2 distance between each pair of nodes.""" # N x N # d_scores = tf.matmul(node_features, tf.transpose(node_features,perm=[1, 0])) # diag = tf.diag_part(d_scores) # d_scores *= -2. # d_scores += tf.reshape(diag, (-1, 1)) + tf.reshape(diag, (1, -1)) l2_norm = tf.reduce_sum(tf.square(node_features), 1) na = tf.reshape(l2_norm, [-1, 1]) nb = tf.reshape(l2_norm, [1, -1]) # return pairwise euclidead difference matrix l2_scores = tf.maximum( na - 2*tf.matmul(node_features, node_features, False, True) + nb, 0.0) return l2_scores def compute_dot_sim_matrix(node_features): """Compute edge scores with dot product.""" sim = tf.matmul(node_features, tf.transpose(node_features, perm=[1, 0])) return sim def compute_dot_norm(features): """Compute edge scores with normalized dot product.""" features = tf.nn.l2_normalize(features, axis=-1) sim = tf.matmul(features, tf.transpose(features, perm=[1, 0])) return sim def compute_asym_dot(node_features): """Compute edge scores with asymmetric dot product.""" feat_left, feat_right = tf.split(node_features, 2, axis=-1) feat_left = tf.nn.l2_normalize(feat_left, axis=-1) feat_right = tf.nn.l2_normalize(feat_right, axis=-1) sim = tf.matmul(feat_left, tf.transpose(feat_right, perm=[1, 0])) return sim def compute_adj(features, att_mechanism, p_drop, is_training): """Compute adj matrix given node features.""" features = tf.layers.dropout( inputs=features, rate=p_drop, training=is_training) if att_mechanism == 'dot': return compute_dot_sim_matrix(features) elif att_mechanism == 'weighted-mat-dot': return compute_weighted_mat_dot(features) elif att_mechanism == 'weighted-dot': return compute_weighted_dot(features) elif att_mechanism == 'att': return compute_adj_att(features) elif att_mechanism == 'dot-norm': return compute_dot_norm(features) elif att_mechanism == 'asym-dot': return compute_asym_dot(features) else: return compute_l2_sim_matrix(features) def get_sp_topk(adj_pred, sp_adj_train, nb_nodes, k): """Returns binary matrix with topK.""" _, indices = tf.nn.top_k(tf.reshape(adj_pred, (-1,)), k) indices = tf.reshape(tf.cast(indices, tf.int64), (-1, 1)) sp_adj_pred = tf.SparseTensor( indices=indices, values=tf.ones(k), dense_shape=(nb_nodes * nb_nodes,)) sp_adj_pred = tf.sparse_reshape(sp_adj_pred, shape=(nb_nodes, nb_nodes, 1)) sp_adj_train = tf.SparseTensor( indices=sp_adj_train.indices, values=tf.ones_like(sp_adj_train.values), dense_shape=sp_adj_train.dense_shape) sp_adj_train = tf.sparse_reshape(sp_adj_train, shape=(nb_nodes, nb_nodes, 1)) sp_adj_pred = tf.sparse_concat( sp_inputs=[sp_adj_pred, sp_adj_train], axis=-1) return tf.sparse_reduce_max(sp_adj_pred, axis=-1) @tf.custom_gradient def mask_edges(scores, mask): masked_scores = tf.multiply(scores, mask) def grad(dy): return dy, None # tf.multiply(scores, dy) return masked_scores, grad
19,241
35.033708
80
py
gcnn-survey-paper
gcnn-survey-paper-master/utils/data_utils.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Utils functions to load and process citation data.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import pickle as pkl import sys import networkx as nx import numpy as np import scipy.sparse as sp from scipy.sparse.linalg.eigen.arpack import eigsh import tensorflow as tf from third_party.gcn.gcn.utils import normalize_adj from third_party.gcn.gcn.utils import parse_index_file from third_party.gcn.gcn.utils import sample_mask from third_party.gcn.gcn.utils import sparse_to_tuple from third_party.gcn.gcn.utils import preprocess_features def load_test_edge_mask(dataset_str, data_path, drop_edge_prop): """Remove test edges by loading edge masks.""" edge_mask_path = os.path.join( data_path, 'emask.{}.remove{}.npz'.format(dataset_str, drop_edge_prop)) with tf.gfile.Open(edge_mask_path) as f: mask = sp.load_npz(f) return mask def load_edge_masks(dataset_str, data_path, adj_true, drop_edge_prop): """Loads adjacency matrix as sparse matrix and masks for val & test links. Args: dataset_str: dataset to use data_path: path to data folder adj_true: true adjacency matrix in dense format, drop_edge_prop: proportion of edges to remove. Returns: adj_matrix: adjacency matrix train_mask: mask for train edges val_mask: mask for val edges test_mask: mask for test edges """ edge_mask_path = os.path.join( data_path, 'emask.{}.remove{}.'.format(dataset_str, drop_edge_prop)) val_mask = sp.load_npz(edge_mask_path + 'val.npz') test_mask = sp.load_npz(edge_mask_path + 'test.npz') train_mask = 1. - val_mask.todense() - test_mask.todense() # remove val and test edges from true A adj_train = np.multiply(adj_true, train_mask) train_mask -= np.eye(train_mask.shape[0]) return adj_train, sparse_to_tuple(val_mask), sparse_to_tuple( val_mask), sparse_to_tuple(test_mask) def add_top_k_edges(data, edge_mask_path, gae_scores_path, topk, nb_nodes, norm_adj): """Loads GAE scores and adds topK edges to train adjacency.""" test_mask = sp.load_npz(os.path.join(edge_mask_path, 'test_mask.npz')) train_mask = 1. - test_mask.todense() # remove val and test edges from true A adj_train_curr = np.multiply(data['adj_true'], train_mask) # Predict test edges using precomputed scores scores = np.load(os.path.join(gae_scores_path, 'gae_scores.npy')) # scores_mask = 1 - np.eye(nb_nodes) scores_mask = np.zeros((nb_nodes, nb_nodes)) scores_mask[:140, 140:] = 1. scores_mask[140:, :140] = 1. scores = np.multiply(scores, scores_mask).reshape((-1,)) threshold = scores[np.argsort(-scores)[topk]] adj_train_curr += 1 * (scores > threshold).reshape((nb_nodes, nb_nodes)) adj_train_curr = 1 * (adj_train_curr > 0) if norm_adj: adj_train_norm = normalize_adj(data['adj_train']) else: adj_train_norm = sp.coo_matrix(data['adj_train']) return adj_train_curr, sparse_to_tuple(adj_train_norm) def process_adj(adj, model_name): """Symmetrically normalize adjacency matrix.""" if model_name == 'Cheby': laplacian = sp.eye(adj.shape[0]) - normalize_adj(adj - sp.eye(adj.shape[0])) # TODO(chamii): compare with # adj) largest_eigval, _ = eigsh(laplacian, 1, which='LM') laplacian_norm = (2. / largest_eigval[0]) * laplacian - sp.eye(adj.shape[0]) return laplacian_norm else: return normalize_adj(adj) def load_data(dataset_str, data_path): if dataset_str in ['cora', 'citeseer', 'pubmed']: return load_citation_data(dataset_str, data_path) else: return load_ppi_data(data_path) def load_ppi_data(data_path): """Load PPI dataset.""" with tf.gfile.Open(os.path.join(data_path, 'ppi.edges.npz')) as f: adj = sp.load_npz(f) with tf.gfile.Open(os.path.join(data_path, 'ppi.features.norm.npy')) as f: features = np.load(f) with tf.gfile.Open(os.path.join(data_path, 'ppi.labels.npz')) as f: labels = sp.load_npz(f).todense() train_mask = np.load( tf.gfile.Open(os.path.join(data_path, 'ppi.train_mask.npy'))) > 0 val_mask = np.load( tf.gfile.Open(os.path.join(data_path, 'ppi.test_mask.npy'))) > 0 test_mask = np.load( tf.gfile.Open(os.path.join(data_path, 'ppi.test_mask.npy'))) > 0 return adj, features, labels, train_mask, val_mask, test_mask def load_citation_data(dataset_str, data_path): """Load data.""" names = ['x', 'y', 'tx', 'ty', 'allx', 'ally', 'graph'] objects = {} for name in names: with tf.gfile.Open( os.path.join(data_path, 'ind.{}.{}'.format(dataset_str, name)), 'rb') as f: if sys.version_info > (3, 0): objects[name] = pkl.load(f) # , encoding='latin1') comment to pass lint else: objects[name] = pkl.load(f) test_idx_reorder = parse_index_file( os.path.join(data_path, 'ind.{}.test.index'.format(dataset_str))) test_idx_range = np.sort(test_idx_reorder) if dataset_str == 'citeseer': # Fix citeseer dataset (there are some isolated nodes in the graph) # Find isolated nodes, add them as zero-vecs into the right position test_idx_range_full = range( min(test_idx_reorder), max(test_idx_reorder) + 1) tx_extended = sp.lil_matrix((len(test_idx_range_full), objects['x'].shape[1])) tx_extended[test_idx_range - min(test_idx_range), :] = objects['tx'] objects['tx'] = tx_extended ty_extended = np.zeros((len(test_idx_range_full), objects['y'].shape[1])) ty_extended[test_idx_range - min(test_idx_range), :] = objects['ty'] objects['ty'] = ty_extended features = sp.vstack((objects['allx'], objects['tx'])).tolil() features[test_idx_reorder, :] = features[test_idx_range, :] adj = nx.adjacency_matrix(nx.from_dict_of_lists(objects['graph'])) labels = np.vstack((objects['ally'], objects['ty'])) labels[test_idx_reorder, :] = labels[test_idx_range, :] idx_test = test_idx_range.tolist() idx_train = range(len(objects['y'])) idx_val = range(len(objects['y']), len(objects['y']) + 500) train_mask = sample_mask(idx_train, labels.shape[0]) val_mask = sample_mask(idx_val, labels.shape[0]) test_mask = sample_mask(idx_test, labels.shape[0]) features = preprocess_features(features) return adj, features, labels, train_mask, val_mask, test_mask def construct_feed_dict(adj_normalized, adj, features, placeholders): # construct feed dictionary feed_dict = dict() feed_dict.update({placeholders['features']: features}) feed_dict.update({placeholders['adj']: adj_normalized}) feed_dict.update({placeholders['adj_orig']: adj}) return feed_dict def mask_val_test_edges(adj, prop): """Function to mask test and val edges.""" # NOTE: Splits are randomized and results might slightly # deviate from reported numbers in the paper. # Remove diagonal elements adj = adj - sp.dia_matrix( (adj.diagonal()[np.newaxis, :], [0]), shape=adj.shape) adj.eliminate_zeros() # Check that diag is zero: assert np.diag(adj.todense()).sum() == 0 adj_triu = sp.triu(adj) adj_tuple = sparse_to_tuple(adj_triu) edges = adj_tuple[0] edges_all = sparse_to_tuple(adj)[0] num_test = int(np.floor(edges.shape[0] * prop)) # num_val = int(np.floor(edges.shape[0] * 0.05)) # we keep 5% for validation # we keep 10% of training edges for validation num_val = int(np.floor((edges.shape[0] - num_test) * 0.05)) all_edge_idx = range(edges.shape[0]) np.random.shuffle(all_edge_idx) val_edge_idx = all_edge_idx[:num_val] test_edge_idx = all_edge_idx[num_val:(num_val + num_test)] test_edges = edges[test_edge_idx] val_edges = edges[val_edge_idx] train_edges = np.delete( edges, np.hstack([test_edge_idx, val_edge_idx]), axis=0) def ismember(a, b, tol=5): rows_close = np.all(np.round(a - b[:, None], tol) == 0, axis=-1) return np.any(rows_close) test_edges_false = [] while len(test_edges_false) < len(test_edges): idx_i = np.random.randint(0, adj.shape[0]) idx_j = np.random.randint(0, adj.shape[0]) if idx_i == idx_j: continue if ismember([idx_i, idx_j], edges_all): continue if test_edges_false: if ismember([idx_j, idx_i], np.array(test_edges_false)): continue if ismember([idx_i, idx_j], np.array(test_edges_false)): continue test_edges_false.append([idx_i, idx_j]) val_edges_false = [] while len(val_edges_false) < len(val_edges): idx_i = np.random.randint(0, adj.shape[0]) idx_j = np.random.randint(0, adj.shape[0]) if idx_i == idx_j: continue if ismember([idx_i, idx_j], train_edges): continue if ismember([idx_j, idx_i], train_edges): continue if ismember([idx_i, idx_j], val_edges): continue if ismember([idx_j, idx_i], val_edges): continue if val_edges_false: if ismember([idx_j, idx_i], np.array(val_edges_false)): continue if ismember([idx_i, idx_j], np.array(val_edges_false)): continue val_edges_false.append([idx_i, idx_j]) assert ~ismember(test_edges_false, edges_all) assert ~ismember(val_edges_false, edges_all) assert ~ismember(val_edges, train_edges) assert ~ismember(test_edges, train_edges) assert ~ismember(val_edges, test_edges) data = np.ones(train_edges.shape[0]) # Re-build adj matrix adj_train = sp.csr_matrix((data, (train_edges[:, 0], train_edges[:, 1])), shape=adj.shape) adj_train = adj_train + adj_train.T # NOTE: these edge lists only contain single direction of edge! num_nodes = adj.shape[0] val_mask = np.zeros((num_nodes, num_nodes)) for i, j in val_edges: val_mask[i, j] = 1 val_mask[j, i] = 1 for i, j in val_edges_false: val_mask[i, j] = 1 val_mask[j, i] = 1 test_mask = np.zeros((num_nodes, num_nodes)) for i, j in test_edges: test_mask[i, j] = 1 test_mask[j, i] = 1 for i, j in test_edges_false: test_mask[i, j] = 1 test_mask[j, i] = 1 return adj_train, sparse_to_tuple(val_mask), sparse_to_tuple(test_mask) def mask_test_edges(adj, prop): """Function to mask test edges. Args: adj: scipy sparse matrix prop: proportion of edges to remove (float in [0, 1]) Returns: adj_train: adjacency with edges removed test_edges: list of positive and negative test edges """ # Remove diagonal elements adj = adj - sp.dia_matrix( (adj.diagonal()[np.newaxis, :], [0]), shape=adj.shape) adj.eliminate_zeros() # Check that diag is zero: assert np.diag(adj.todense()).sum() == 0 adj_triu = sp.triu(adj) adj_tuple = sparse_to_tuple(adj_triu) edges = adj_tuple[0] edges_all = sparse_to_tuple(adj)[0] num_test = int(np.floor(edges.shape[0] * prop)) all_edge_idx = range(edges.shape[0]) np.random.shuffle(all_edge_idx) test_edge_idx = all_edge_idx[:num_test] test_edges = edges[test_edge_idx] train_edges = np.delete(edges, test_edge_idx, axis=0) def ismember(a, b, tol=5): rows_close = np.all(np.round(a - b[:, None], tol) == 0, axis=-1) return np.any(rows_close) test_edges_false = [] while len(test_edges_false) < len(test_edges): idx_i = np.random.randint(0, adj.shape[0]) idx_j = np.random.randint(0, adj.shape[0]) if idx_i == idx_j: continue if ismember([idx_i, idx_j], edges_all): continue if test_edges_false: if ismember([idx_j, idx_i], np.array(test_edges_false)): continue if ismember([idx_i, idx_j], np.array(test_edges_false)): continue test_edges_false.append([idx_i, idx_j]) assert ~ismember(test_edges_false, edges_all) assert ~ismember(test_edges, train_edges) data = np.ones(train_edges.shape[0]) # Re-build adj matrix adj_train = sp.csr_matrix((data, (train_edges[:, 0], train_edges[:, 1])), shape=adj.shape) adj_train = adj_train + adj_train.T # NOTE: these edge lists only contain single direction of edge! num_nodes = adj.shape[0] test_mask = np.zeros((num_nodes, num_nodes)) for i, j in test_edges: test_mask[i, j] = 1 test_mask[j, i] = 1 for i, j in test_edges_false: test_mask[i, j] = 1 test_mask[j, i] = 1 return adj_train, sparse_to_tuple(test_mask)
12,831
33.775068
80
py
gcnn-survey-paper
gcnn-survey-paper-master/utils/link_prediction_utils.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Heuristics for link prediction.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from absl import flags from data_utils import mask_test_edges import networkx as nx import numpy as np import scipy.sparse as sp import sklearn.metrics as skm flags.DEFINE_string('adj_path', '../data/cora.adj.npz', 'path to graph to use.') flags.DEFINE_string('prop_drop', '10-30-50', 'proportion of edges to remove.') flags.DEFINE_string('methods', 'svd-katz-common_neighbours', 'which methods to use') FLAGS = flags.FLAGS class LinkPredictionHeuristcs(object): """Link prediction heuristics.""" def __init__(self, adj_matrix): self.adj_matrix = adj_matrix def common_neighbours(self): """Computes scores for each node pair based on common neighbours.""" scores = self.adj_matrix.dot(self.adj_matrix) return scores def svd(self, rank=64): """Computes scores using low rank factorization with SVD.""" adj_matrix = self.adj_matrix.asfptype() u, s, v = sp.linalg.svds(A=adj_matrix, k=rank) adj_low_rank = u.dot(np.diag(s).dot(v)) return adj_low_rank def adamic_adar(self): """Computes adamic adar scores.""" graph = nx.from_scipy_sparse_matrix(self.adj_matrix) scores = nx.adamic_adar_index(graph) return scores def jaccard_coeff(self): """Computes Jaccard coefficients.""" graph = nx.from_scipy_sparse_matrix(self.adj_matrix) coeffs = nx.jaccard_coefficient(graph) return coeffs def katz(self, beta=0.001, steps=25): """Computes Katz scores.""" coeff = beta katz_scores = beta * self.adj_matrix adj_power = self.adj_matrix for _ in range(2, steps + 1): adj_power = adj_power.dot(self.adj_matrix) katz_scores += coeff * adj_power coeff *= beta return katz_scores def get_scores_from_generator(gen, nb_nodes=2708): """Helper function to get scores in numpy array format from generator.""" adj = np.zeros((nb_nodes, nb_nodes)) for i, j, score in gen: adj[i, j] = score return adj def compute_lp_metrics(edges, true_adj, pred_adj): """Computes link prediction scores on test edges.""" labels = np.array(true_adj[edges]).reshape((-1,)) scores = np.array(pred_adj[edges]).reshape((-1,)) roc = skm.roc_auc_score(labels, scores) ap = skm.average_precision_score(labels, scores) return roc, ap if __name__ == '__main__': adj_true = sp.load_npz(FLAGS.adj_path).todense() lp = LinkPredictionHeuristcs(adj_true) for delete_prop in FLAGS.prop_drop.split('-'): for method in FLAGS.methods.split('-'): lp_func = getattr(lp, method) adj_train, test_edges = mask_test_edges( adj_true, float(delete_prop) * 0.01) adj_scores = lp_func(adj_train).todense() roc_score, ap_score = compute_lp_metrics(test_edges, adj_true, adj_scores) print('method={} | prop={} | roc_auc={} ap={}\n'.format( method, delete_prop, round(roc_score, 4), round(ap_score, 4)))
3,591
32.259259
80
py
gcnn-survey-paper
gcnn-survey-paper-master/utils/train_utils.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License. """Helper functions for training.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function def format_metrics(metrics, mode): """Format metrics for logging.""" result = '' for metric in metrics: result += '{}_{} = {:.4f} | '.format(mode, metric, float(metrics[metric])) return result def format_params(config): """Format training parameters for logging.""" result = '' for key, value in config.__dict__.items(): result += '{}={} \n '.format(key, str(value)) return result def check_improve(best_metrics, metrics, targets): """Checks if any of the target metrics improved.""" return [ compare(metrics[target], best_metrics[target], targets[target]) for target in targets ] def compare(x1, x2, increasing): if increasing == 1: return x1 >= x2 else: return x1 <= x2
1,447
27.392157
78
py
gcnn-survey-paper
gcnn-survey-paper-master/utils/__init__.py
#Copyright 2018 Google LLC # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # #Unless required by applicable law or agreed to in writing, software #distributed under the License is distributed on an "AS IS" BASIS, #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #See the License for the specific language governing permissions and #limitations under the License.
565
39.428571
73
py
nervaluate
nervaluate-main/examples/example_no_loader.py
import nltk import sklearn_crfsuite from sklearn.metrics import classification_report from nervaluate import collect_named_entities, summary_report_ent, summary_report_overall from nervaluate import Evaluator def word2features(sent, i): word = sent[i][0] postag = sent[i][1] features = { "bias": 1.0, "word.lower()": word.lower(), "word[-3:]": word[-3:], "word[-2:]": word[-2:], "word.isupper()": word.isupper(), "word.istitle()": word.istitle(), "word.isdigit()": word.isdigit(), "postag": postag, "postag[:2]": postag[:2], } if i > 0: word1 = sent[i - 1][0] postag1 = sent[i - 1][1] features.update( { "-1:word.lower()": word1.lower(), "-1:word.istitle()": word1.istitle(), "-1:word.isupper()": word1.isupper(), "-1:postag": postag1, "-1:postag[:2]": postag1[:2], } ) else: features["BOS"] = True if i < len(sent) - 1: word1 = sent[i + 1][0] postag1 = sent[i + 1][1] features.update( { "+1:word.lower()": word1.lower(), "+1:word.istitle()": word1.istitle(), "+1:word.isupper()": word1.isupper(), "+1:postag": postag1, "+1:postag[:2]": postag1[:2], } ) else: features["EOS"] = True return features def sent2features(sent): return [word2features(sent, i) for i in range(len(sent))] def sent2labels(sent): return [label for token, postag, label in sent] def sent2tokens(sent): return [token for token, postag, label in sent] def main(): print("Loading CoNLL 2002 NER Spanish data") nltk.corpus.conll2002.fileids() train_sents = list(nltk.corpus.conll2002.iob_sents("esp.train")) test_sents = list(nltk.corpus.conll2002.iob_sents("esp.testb")) x_train = [sent2features(s) for s in train_sents] y_train = [sent2labels(s) for s in train_sents] x_test = [sent2features(s) for s in test_sents] y_test = [sent2labels(s) for s in test_sents] print("Train a CRF on the CoNLL 2002 NER Spanish data") crf = sklearn_crfsuite.CRF(algorithm="lbfgs", c1=0.1, c2=0.1, max_iterations=10, all_possible_transitions=True) try: crf.fit(x_train, y_train) except AttributeError: pass y_pred = crf.predict(x_test) labels = list(crf.classes_) labels.remove("O") # remove 'O' label from evaluation sorted_labels = sorted(labels, key=lambda name: (name[1:], name[0])) # group B- and I- results y_test_flat = [y for msg in y_test for y in msg] y_pred_flat = [y for msg in y_pred for y in msg] print(classification_report(y_test_flat, y_pred_flat, labels=sorted_labels)) test_sents_labels = [] for sentence in test_sents: sentence = [token[2] for token in sentence] test_sents_labels.append(sentence) pred_collected = [collect_named_entities(msg) for msg in y_pred] test_collected = [collect_named_entities(msg) for msg in y_test] evaluator = Evaluator(test_collected, pred_collected, ["LOC", "MISC", "PER", "ORG"]) results, results_agg = evaluator.evaluate() print("\n\nOverall") print(summary_report_overall(results)) print("\n\n'Strict'") print(summary_report_ent(results_agg, scenario="strict")) print("\n\n'Ent_Type'") print(summary_report_ent(results_agg, scenario="ent_type")) print("\n\n'Partial'") print(summary_report_ent(results_agg, scenario="partial")) print("\n\n'Exact'") print(summary_report_ent(results_agg, scenario="exact")) if __name__ == "__main__": main()
3,759
30.07438
115
py
nervaluate
nervaluate-main/src/nervaluate/evaluate.py
import logging from copy import deepcopy from typing import List, Dict, Union, Tuple from .utils import conll_to_spans, find_overlap, list_to_spans class Evaluator: # pylint: disable=too-many-instance-attributes, too-few-public-methods def __init__( self, true: Union[List[List[str]], List[str], List[Dict], str, List[List[Dict[str, Union[int, str]]]]], pred: Union[List[List[str]], List[str], List[Dict], str, List[List[Dict[str, Union[int, str]]]]], tags: List[str], loader: str = "default", ) -> None: self.true = true self.pred = pred self.tags = tags # self.list = [] # Setup dict into which metrics will be stored. self.metrics_results = { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 0, "actual": 0, "precision": 0, "recall": 0, "f1": 0, } # Copy results dict to cover the four schemes. self.results = { "strict": deepcopy(self.metrics_results), "ent_type": deepcopy(self.metrics_results), "partial": deepcopy(self.metrics_results), "exact": deepcopy(self.metrics_results), } # Create an accumulator to store results self.evaluation_agg_entities_type = {e: deepcopy(self.results) for e in tags} self.loaders = { "list": list_to_spans, "conll": conll_to_spans, } self.loader = loader def evaluate(self) -> Tuple[Dict, Dict]: logging.debug("Imported %s predictions for %s true examples", len(self.pred), len(self.true)) if self.loader != "default": loader = self.loaders[self.loader] self.pred = loader(self.pred) self.true = loader(self.true) if len(self.true) != len(self.pred): raise ValueError("Number of predicted documents does not equal true") for true_ents, pred_ents in zip(self.true, self.pred): # Compute results for one message tmp_results, tmp_agg_results = compute_metrics(true_ents, pred_ents, self.tags) # Cycle through each result and accumulate # TODO: Combine these loops below: for eval_schema in self.results: for metric in self.results[eval_schema]: self.results[eval_schema][metric] += tmp_results[eval_schema][metric] # Calculate global precision and recall self.results = compute_precision_recall_wrapper(self.results) # Aggregate results by entity type for label in self.tags: for eval_schema in tmp_agg_results[label]: for metric in tmp_agg_results[label][eval_schema]: self.evaluation_agg_entities_type[label][eval_schema][metric] += tmp_agg_results[label][ eval_schema ][metric] # Calculate precision recall at the individual entity level self.evaluation_agg_entities_type[label] = compute_precision_recall_wrapper( self.evaluation_agg_entities_type[label] ) return self.results, self.evaluation_agg_entities_type # flake8: noqa: C901 def compute_metrics( # type: ignore true_named_entities, pred_named_entities, tags: List[str] ): # pylint: disable=too-many-locals, too-many-branches, too-many-statements """ Compute metrics on the collected true and predicted named entities :true_name_entities: collected true named entities output by collect_named_entities :pred_name_entities: collected predicted named entities output by collect_named_entities :tags: list of tags to be used """ eval_metrics = { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "precision": 0, "recall": 0, "f1": 0, } # overall results evaluation = { "strict": deepcopy(eval_metrics), "ent_type": deepcopy(eval_metrics), "partial": deepcopy(eval_metrics), "exact": deepcopy(eval_metrics), } # results by entity type evaluation_agg_entities_type = {e: deepcopy(evaluation) for e in tags} # keep track of entities that overlapped true_which_overlapped_with_pred = [] # Subset into only the tags that we are interested in. # NOTE: we remove the tags we don't want from both the predicted and the # true entities. This covers the two cases where mismatches can occur: # # 1) Where the model predicts a tag that is not present in the true data # 2) Where there is a tag in the true data that the model is not capable of # predicting. # Strip the spans down to just start, end, label. Note that failing # to do this results in a bug. The exact cause is not clear. true_named_entities = [clean_entities(ent) for ent in true_named_entities if ent["label"] in tags] pred_named_entities = [clean_entities(ent) for ent in pred_named_entities if ent["label"] in tags] # go through each predicted named-entity for pred in pred_named_entities: found_overlap = False # Check each of the potential scenarios in turn. See # http://www.davidsbatista.net/blog/2018/05/09/Named_Entity_Evaluation/ # for scenario explanation. # Scenario I: Exact match between true and pred if pred in true_named_entities: true_which_overlapped_with_pred.append(pred) evaluation["strict"]["correct"] += 1 evaluation["ent_type"]["correct"] += 1 evaluation["exact"]["correct"] += 1 evaluation["partial"]["correct"] += 1 # for the agg. by label results evaluation_agg_entities_type[pred["label"]]["strict"]["correct"] += 1 evaluation_agg_entities_type[pred["label"]]["ent_type"]["correct"] += 1 evaluation_agg_entities_type[pred["label"]]["exact"]["correct"] += 1 evaluation_agg_entities_type[pred["label"]]["partial"]["correct"] += 1 else: # check for overlaps with any of the true entities for true in true_named_entities: # overlapping needs to take into account last token as well pred_range = range(pred["start"], pred["end"] + 1) true_range = range(true["start"], true["end"] + 1) # Scenario IV: Offsets match, but entity type is wrong if true["start"] == pred["start"] and pred["end"] == true["end"] and true["label"] != pred["label"]: # overall results evaluation["strict"]["incorrect"] += 1 evaluation["ent_type"]["incorrect"] += 1 evaluation["partial"]["correct"] += 1 evaluation["exact"]["correct"] += 1 # aggregated by entity type results evaluation_agg_entities_type[true["label"]]["strict"]["incorrect"] += 1 evaluation_agg_entities_type[true["label"]]["ent_type"]["incorrect"] += 1 evaluation_agg_entities_type[true["label"]]["partial"]["correct"] += 1 evaluation_agg_entities_type[true["label"]]["exact"]["correct"] += 1 true_which_overlapped_with_pred.append(true) found_overlap = True break # check for an overlap i.e. not exact boundary match, with true entities # overlaps with true entities must only count once if find_overlap(true_range, pred_range) and true not in true_which_overlapped_with_pred: true_which_overlapped_with_pred.append(true) # Scenario V: There is an overlap (but offsets do not match # exactly), and the entity type is the same. # 2.1 overlaps with the same entity type if pred["label"] == true["label"]: # overall results evaluation["strict"]["incorrect"] += 1 evaluation["ent_type"]["correct"] += 1 evaluation["partial"]["partial"] += 1 evaluation["exact"]["incorrect"] += 1 # aggregated by entity type results evaluation_agg_entities_type[true["label"]]["strict"]["incorrect"] += 1 evaluation_agg_entities_type[true["label"]]["ent_type"]["correct"] += 1 evaluation_agg_entities_type[true["label"]]["partial"]["partial"] += 1 evaluation_agg_entities_type[true["label"]]["exact"]["incorrect"] += 1 found_overlap = True break # Scenario VI: Entities overlap, but the entity type is # different. # overall results evaluation["strict"]["incorrect"] += 1 evaluation["ent_type"]["incorrect"] += 1 evaluation["partial"]["partial"] += 1 evaluation["exact"]["incorrect"] += 1 # aggregated by entity type results # Results against the true entity evaluation_agg_entities_type[true["label"]]["strict"]["incorrect"] += 1 evaluation_agg_entities_type[true["label"]]["partial"]["partial"] += 1 evaluation_agg_entities_type[true["label"]]["ent_type"]["incorrect"] += 1 evaluation_agg_entities_type[true["label"]]["exact"]["incorrect"] += 1 # Results against the predicted entity # evaluation_agg_entities_type[pred['label']]['strict']['spurious'] += 1 found_overlap = True break # Scenario II: Entities are spurious (i.e., over-generated). if not found_overlap: # Overall results evaluation["strict"]["spurious"] += 1 evaluation["ent_type"]["spurious"] += 1 evaluation["partial"]["spurious"] += 1 evaluation["exact"]["spurious"] += 1 # Aggregated by entity type results # a over-generated entity with a valid tag should be # attributed to the respective tag only if pred["label"] in tags: spurious_tags = [pred["label"]] else: # NOTE: when pred.e_type is not found in valid tags # or when it simply does not appear in the test set, then it is # spurious, but it is not clear where to assign it at the tag # level. In this case, it is applied to all target_tags # found in this example. This will mean that the sum of the # evaluation_agg_entities will not equal evaluation. spurious_tags = tags for true in spurious_tags: evaluation_agg_entities_type[true]["strict"]["spurious"] += 1 evaluation_agg_entities_type[true]["ent_type"]["spurious"] += 1 evaluation_agg_entities_type[true]["partial"]["spurious"] += 1 evaluation_agg_entities_type[true]["exact"]["spurious"] += 1 # Scenario III: Entity was missed entirely. for true in true_named_entities: if true in true_which_overlapped_with_pred: continue # overall results evaluation["strict"]["missed"] += 1 evaluation["ent_type"]["missed"] += 1 evaluation["partial"]["missed"] += 1 evaluation["exact"]["missed"] += 1 # for the agg. by label evaluation_agg_entities_type[true["label"]]["strict"]["missed"] += 1 evaluation_agg_entities_type[true["label"]]["ent_type"]["missed"] += 1 evaluation_agg_entities_type[true["label"]]["partial"]["missed"] += 1 evaluation_agg_entities_type[true["label"]]["exact"]["missed"] += 1 # Compute 'possible', 'actual' according to SemEval-2013 Task 9.1 on the # overall results, and use these to calculate precision and recall. for eval_type in evaluation: evaluation[eval_type] = compute_actual_possible(evaluation[eval_type]) # Compute 'possible', 'actual', and precision and recall on entity level # results. Start by cycling through the accumulated results. for entity_type, entity_level in evaluation_agg_entities_type.items(): # Cycle through the evaluation types for each dict containing entity # level results. for eval_type in entity_level: evaluation_agg_entities_type[entity_type][eval_type] = compute_actual_possible(entity_level[eval_type]) return evaluation, evaluation_agg_entities_type def compute_actual_possible(results: Dict) -> Dict: """ Takes a result dict that has been output by compute metrics. Returns the results' dict with actual, possible populated. When the results dicts is from partial or ent_type metrics, then partial_or_type=True to ensure the right calculation is used for calculating precision and recall. """ correct = results["correct"] incorrect = results["incorrect"] partial = results["partial"] missed = results["missed"] spurious = results["spurious"] # Possible: number annotations in the gold-standard which contribute to the # final score possible = correct + incorrect + partial + missed # Actual: number of annotations produced by the NER system actual = correct + incorrect + partial + spurious results["actual"] = actual results["possible"] = possible return results def compute_precision_recall(results: Dict, partial_or_type: bool = False) -> Dict: """ Takes a result dict that has been output by compute metrics. Returns the results' dict with precision and recall populated. When the results dicts is from partial or ent_type metrics, then partial_or_type=True to ensure the right calculation is used for calculating precision and recall. """ actual = results["actual"] possible = results["possible"] partial = results["partial"] correct = results["correct"] if partial_or_type: precision = (correct + 0.5 * partial) / actual if actual > 0 else 0 recall = (correct + 0.5 * partial) / possible if possible > 0 else 0 else: precision = correct / actual if actual > 0 else 0 recall = correct / possible if possible > 0 else 0 results["precision"] = precision results["recall"] = recall results["f1"] = 2 * (precision * recall) / (precision + recall) if (precision + recall) > 0 else 0 return results def compute_precision_recall_wrapper(results: Dict) -> Dict: """ Wraps the compute_precision_recall function and runs on a dict of results """ results_a = { key: compute_precision_recall(value, True) for key, value in results.items() if key in ["partial", "ent_type"] } results_b = {key: compute_precision_recall(value) for key, value in results.items() if key in ["strict", "exact"]} results = {**results_a, **results_b} return results def clean_entities(ent: Dict) -> Dict: """ Returns just the useful keys if additional keys are present in the entity dict. This may happen if passing a list of spans directly from prodigy, which typically may include 'token_start' and 'token_end'. """ return {"start": ent["start"], "end": ent["end"], "label": ent["label"]} def summary_report_ent( # pylint: disable=too-many-locals results_agg_entities_type: Dict, scenario: str = "strict", digits: int = 2 ) -> str: if scenario not in {"strict", "ent_type", "partial", "exact"}: raise Exception("Invalid scenario: must be one of 'strict', 'ent_type', 'partial', 'exact'") target_names = sorted(results_agg_entities_type.keys()) headers = ["correct", "incorrect", "partial", "missed", "spurious", "precision", "recall", "f1-score"] rows = [headers] for ent_type, results in sorted(results_agg_entities_type.items()): for k, v in results.items(): if k != scenario: continue rows.append( [ ent_type, v["correct"], v["incorrect"], v["partial"], v["missed"], v["spurious"], v["precision"], v["recall"], v["f1"], ] ) name_width = max(len(cn) for cn in target_names) width = max(name_width, digits) head_fmt = "{:>{width}s} " + " {:>11}" * len(headers) report = head_fmt.format("", *headers, width=width) report += "\n\n" row_fmt = "{:>{width}s} " + " {:>11}" * 5 + " {:>11.{digits}f}" * 3 + "\n" for row in rows[1:]: report += row_fmt.format(*row, width=width, digits=digits) return report def summary_report_overall(results: Dict, digits: int = 2) -> str: headers = ["correct", "incorrect", "partial", "missed", "spurious", "precision", "recall", "f1-score"] rows = [headers] for k, v in results.items(): rows.append( [ k, v["correct"], v["incorrect"], v["partial"], v["missed"], v["spurious"], v["precision"], v["recall"], v["f1"], ] ) target_names = sorted(results.keys()) name_width = max(len(cn) for cn in target_names) width = max(name_width, digits) head_fmt = "{:>{width}s} " + " {:>11}" * len(headers) report = head_fmt.format("", *headers, width=width) report += "\n\n" row_fmt = "{:>{width}s} " + " {:>11}" * 5 + " {:>11.{digits}f}" * 3 + "\n" for row in rows[1:]: report += row_fmt.format(*row, width=width, digits=digits) return report
18,400
38.829004
118
py
nervaluate
nervaluate-main/src/nervaluate/utils.py
from typing import List, Dict, Optional def split_list(token: List[str], split_chars: Optional[List[str]] = None) -> List[List[str]]: if split_chars is None: split_chars = [""] out = [] chunk = [] for i, item in enumerate(token): if item not in split_chars: chunk.append(item) if i + 1 == len(token): out.append(chunk) else: out.append(chunk) chunk = [] return out def conll_to_spans(doc: str) -> List[List[Dict]]: out = [] doc_parts = split_list(doc.split("\n"), split_chars=None) for example in doc_parts: labels = [] for token in example: token_parts = token.split("\t") label = token_parts[1] labels.append(label) out.append(labels) spans = list_to_spans(out) return spans def list_to_spans(doc): # type: ignore spans = [collect_named_entities(tokens) for tokens in doc] return spans def collect_named_entities(tokens: List[str]) -> List[Dict]: """ Creates a list of Entity named-tuples, storing the entity type and the start and end offsets of the entity. :param tokens: a list of tags :return: a list of Entity named-tuples """ named_entities = [] start_offset = None end_offset = None ent_type = None for offset, token_tag in enumerate(tokens): if token_tag == "O": if ent_type is not None and start_offset is not None: end_offset = offset - 1 named_entities.append({"label": ent_type, "start": start_offset, "end": end_offset}) start_offset = None end_offset = None ent_type = None elif ent_type is None: ent_type = token_tag[2:] start_offset = offset elif ent_type != token_tag[2:] or (ent_type == token_tag[2:] and token_tag[:1] == "B"): end_offset = offset - 1 named_entities.append({"label": ent_type, "start": start_offset, "end": end_offset}) # start of a new entity ent_type = token_tag[2:] start_offset = offset end_offset = None # Catches an entity that goes up until the last token if ent_type is not None and start_offset is not None and end_offset is None: named_entities.append({"label": ent_type, "start": start_offset, "end": len(tokens) - 1}) return named_entities def find_overlap(true_range: range, pred_range: range) -> set: """Find the overlap between two ranges Find the overlap between two ranges. Return the overlapping values if present, else return an empty set(). Examples: >>> find_overlap((1, 2), (2, 3)) 2 >>> find_overlap((1, 2), (3, 4)) set() """ true_set = set(true_range) pred_set = set(pred_range) overlaps = true_set.intersection(pred_set) return overlaps
2,951
27.114286
100
py
nervaluate
nervaluate-main/src/nervaluate/__init__.py
from .evaluate import ( Evaluator, compute_actual_possible, compute_metrics, compute_precision_recall, compute_precision_recall_wrapper, find_overlap, summary_report_ent, summary_report_overall, ) from .utils import collect_named_entities, conll_to_spans, list_to_spans, split_list
314
25.25
84
py
nervaluate
nervaluate-main/tests/test_evaluator.py
from nervaluate import Evaluator def test_evaluator_simple_case(): true = [ [{"label": "PER", "start": 2, "end": 4}], [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], ] pred = [ [{"label": "PER", "start": 2, "end": 4}], [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], ] evaluator = Evaluator(true, pred, tags=["LOC", "PER"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "ent_type": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "partial": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "exact": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_evaluator_simple_case_filtered_tags(): """Check that tags can be excluded by passing the tags argument""" true = [ [{"label": "PER", "start": 2, "end": 4}], [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], ] pred = [ [{"label": "PER", "start": 2, "end": 4}], [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], ] evaluator = Evaluator(true, pred, tags=["PER", "LOC"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "ent_type": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "partial": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "exact": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_evaluator_extra_classes(): """Case when model predicts a class that is not in the gold (true) data""" true = [ [{"label": "ORG", "start": 1, "end": 3}], ] pred = [ [{"label": "FOO", "start": 1, "end": 3}], ] evaluator = Evaluator(true, pred, tags=["ORG", "FOO"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 0, "recall": 0.0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 0, "recall": 0.0, "f1": 0, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_evaluator_no_entities_in_prediction(): """Case when model predicts a class that is not in the gold (true) data""" true = [ [{"label": "PER", "start": 2, "end": 4}], ] pred = [ [], ] evaluator = Evaluator(true, pred, tags=["PER"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "possible": 1, "actual": 0, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "possible": 1, "actual": 0, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "possible": 1, "actual": 0, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "possible": 1, "actual": 0, "precision": 0, "recall": 0, "f1": 0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_evaluator_compare_results_and_results_agg(): """Check that the label level results match the total results.""" true = [ [{"label": "PER", "start": 2, "end": 4}], ] pred = [ [{"label": "PER", "start": 2, "end": 4}], ] evaluator = Evaluator(true, pred, tags=["PER"]) results, results_agg = evaluator.evaluate() expected = { "strict": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "ent_type": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, } expected_agg = { "PER": { "strict": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "ent_type": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, } } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] assert results["strict"] == expected_agg["PER"]["strict"] assert results["ent_type"] == expected_agg["PER"]["ent_type"] assert results["partial"] == expected_agg["PER"]["partial"] assert results["exact"] == expected_agg["PER"]["exact"] def test_evaluator_compare_results_and_results_agg_1(): """Test case when model predicts a label not in the test data.""" true = [ [], [{"label": "ORG", "start": 2, "end": 4}], [{"label": "MISC", "start": 2, "end": 4}], ] pred = [ [{"label": "PER", "start": 2, "end": 4}], [{"label": "ORG", "start": 2, "end": 4}], [{"label": "MISC", "start": 2, "end": 4}], ] evaluator = Evaluator(true, pred, tags=["PER", "ORG", "MISC"]) results, results_agg = evaluator.evaluate() expected = { "strict": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.6666666666666666, "recall": 1.0, "f1": 0.8, }, "ent_type": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.6666666666666666, "recall": 1.0, "f1": 0.8, }, "partial": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.6666666666666666, "recall": 1.0, "f1": 0.8, }, "exact": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.6666666666666666, "recall": 1.0, "f1": 0.8, }, } expected_agg = { "ORG": { "strict": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1.0, "recall": 1, "f1": 1.0, }, "ent_type": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1.0, "recall": 1, "f1": 1.0, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1.0, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, }, "MISC": { "strict": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "ent_type": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 1, "actual": 1, "precision": 1, "recall": 1, "f1": 1, }, }, } assert results_agg["ORG"]["strict"] == expected_agg["ORG"]["strict"] assert results_agg["ORG"]["ent_type"] == expected_agg["ORG"]["ent_type"] assert results_agg["ORG"]["partial"] == expected_agg["ORG"]["partial"] assert results_agg["ORG"]["exact"] == expected_agg["ORG"]["exact"] assert results_agg["MISC"]["strict"] == expected_agg["MISC"]["strict"] assert results_agg["MISC"]["ent_type"] == expected_agg["MISC"]["ent_type"] assert results_agg["MISC"]["partial"] == expected_agg["MISC"]["partial"] assert results_agg["MISC"]["exact"] == expected_agg["MISC"]["exact"] assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_evaluator_with_extra_keys_in_pred(): true = [ [{"label": "PER", "start": 2, "end": 4}], [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], ] pred = [ [{"label": "PER", "start": 2, "end": 4, "token_start": 0, "token_end": 5}], [ {"label": "LOC", "start": 1, "end": 2, "token_start": 0, "token_end": 6}, {"label": "LOC", "start": 3, "end": 4, "token_start": 0, "token_end": 3}, ], ] evaluator = Evaluator(true, pred, tags=["LOC", "PER"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "ent_type": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "partial": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "exact": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_evaluator_with_extra_keys_in_true(): true = [ [{"label": "PER", "start": 2, "end": 4, "token_start": 0, "token_end": 4}], [ {"label": "LOC", "start": 1, "end": 2, "token_start": 0, "token_end": 5}, {"label": "LOC", "start": 3, "end": 4, "token_start": 7, "token_end": 9}, ], ] pred = [ [{"label": "PER", "start": 2, "end": 4}], [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], ] evaluator = Evaluator(true, pred, tags=["LOC", "PER"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "ent_type": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "partial": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, "exact": { "correct": 3, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "possible": 3, "actual": 3, "precision": 1.0, "recall": 1.0, "f1": 1.0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_issue_29(): true = [ [ {"label": "PER", "start": 1, "end": 2}, {"label": "PER", "start": 3, "end": 10}, ] ] pred = [ [ {"label": "PER", "start": 1, "end": 2}, {"label": "PER", "start": 3, "end": 5}, {"label": "PER", "start": 6, "end": 10}, ] ] evaluator = Evaluator(true, pred, tags=["PER"]) results, _ = evaluator.evaluate() expected = { "strict": { "correct": 1, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.3333333333333333, "recall": 0.5, "f1": 0.4, }, "ent_type": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.6666666666666666, "recall": 1.0, "f1": 0.8, }, "partial": { "correct": 1, "incorrect": 0, "partial": 1, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.5, "recall": 0.75, "f1": 0.6, }, "exact": { "correct": 1, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 1, "possible": 2, "actual": 3, "precision": 0.3333333333333333, "recall": 0.5, "f1": 0.4, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"]
22,589
26.616137
85
py
nervaluate
nervaluate-main/tests/test_loaders.py
from nervaluate import Evaluator def test_loaders_produce_the_same_results(): true_list = [ ["O", "O", "O", "O", "O", "O"], ["O", "O", "B-ORG", "I-ORG", "O", "O"], ["O", "O", "B-MISC", "I-MISC", "O", "O"], ["B-MISC", "I-MISC", "I-MISC", "I-MISC", "I-MISC", "I-MISC"], ] pred_list = [ ["O", "O", "B-PER", "I-PER", "O", "O"], ["O", "O", "B-ORG", "I-ORG", "O", "O"], ["O", "O", "B-MISC", "I-MISC", "O", "O"], ["B-MISC", "I-MISC", "I-MISC", "I-MISC", "I-MISC", "I-MISC"], ] true_conll = ( "word\tO\nword\tO\nword\tO\nword\tO\nword\tO\nword\tO\n\n" "word\tO\nword\tO\nword\tB-ORG\nword\tI-ORG\nword\tO\nword\tO\n\n" "word\tO\nword\tO\nword\tB-MISC\nword\tI-MISC\nword\tO\nword\tO\n\n" "word\tB-MISC\nword\tI-MISC\nword\tI-MISC\nword\tI-MISC\nword\tI-MISC\nword\tI-MISC\n" ) pred_conll = ( "word\tO\nword\tO\nword\tB-PER\nword\tI-PER\nword\tO\nword\tO\n\n" "word\tO\nword\tO\nword\tB-ORG\nword\tI-ORG\nword\tO\nword\tO\n\n" "word\tO\nword\tO\nword\tB-MISC\nword\tI-MISC\nword\tO\nword\tO\n\n" "word\tB-MISC\nword\tI-MISC\nword\tI-MISC\nword\tI-MISC\nword\tI-MISC\nword\tI-MISC\n" ) true_prod = [ [], [{"label": "ORG", "start": 2, "end": 3}], [{"label": "MISC", "start": 2, "end": 3}], [{"label": "MISC", "start": 0, "end": 5}], ] pred_prod = [ [{"label": "PER", "start": 2, "end": 3}], [{"label": "ORG", "start": 2, "end": 3}], [{"label": "MISC", "start": 2, "end": 3}], [{"label": "MISC", "start": 0, "end": 5}], ] evaluator_list = Evaluator(true_list, pred_list, tags=["PER", "ORG", "MISC"], loader="list") evaluator_conll = Evaluator(true_conll, pred_conll, tags=["PER", "ORG", "MISC"], loader="conll") evaluator_prod = Evaluator(true_prod, pred_prod, tags=["PER", "ORG", "MISC"]) _, _ = evaluator_list.evaluate() _, _ = evaluator_prod.evaluate() _, _ = evaluator_conll.evaluate() assert evaluator_prod.pred == evaluator_list.pred == evaluator_conll.pred assert evaluator_prod.true == evaluator_list.true == evaluator_conll.true
2,210
36.474576
100
py
nervaluate
nervaluate-main/tests/__init__.py
import sys sys.path.append("../src/nervaluate")
49
11.5
36
py
nervaluate
nervaluate-main/tests/test_nervaluate.py
from nervaluate import ( compute_actual_possible, compute_metrics, compute_precision_recall, compute_precision_recall_wrapper, ) def test_compute_metrics_case_1(): true_named_entities = [ {"label": "PER", "start": 59, "end": 69}, {"label": "LOC", "start": 127, "end": 134}, {"label": "LOC", "start": 164, "end": 174}, {"label": "LOC", "start": 197, "end": 205}, {"label": "LOC", "start": 208, "end": 219}, {"label": "MISC", "start": 230, "end": 240}, ] pred_named_entities = [ {"label": "PER", "start": 24, "end": 30}, {"label": "LOC", "start": 124, "end": 134}, {"label": "PER", "start": 164, "end": 174}, {"label": "LOC", "start": 197, "end": 205}, {"label": "LOC", "start": 208, "end": 219}, {"label": "LOC", "start": 225, "end": 243}, ] results, _ = compute_metrics(true_named_entities, pred_named_entities, ["PER", "LOC", "MISC"]) results = compute_precision_recall_wrapper(results) expected = { "strict": { "correct": 2, "incorrect": 3, "partial": 0, "missed": 1, "spurious": 1, "possible": 6, "actual": 6, "precision": 0.3333333333333333, "recall": 0.3333333333333333, "f1": 0.3333333333333333, }, "ent_type": { "correct": 3, "incorrect": 2, "partial": 0, "missed": 1, "spurious": 1, "possible": 6, "actual": 6, "precision": 0.5, "recall": 0.5, "f1": 0.5, }, "partial": { "correct": 3, "incorrect": 0, "partial": 2, "missed": 1, "spurious": 1, "possible": 6, "actual": 6, "precision": 0.6666666666666666, "recall": 0.6666666666666666, "f1": 0.6666666666666666, }, "exact": { "correct": 3, "incorrect": 2, "partial": 0, "missed": 1, "spurious": 1, "possible": 6, "actual": 6, "precision": 0.5, "recall": 0.5, "f1": 0.5, }, } assert results == expected def test_compute_metrics_agg_scenario_3(): true_named_entities = [{"label": "PER", "start": 59, "end": 69}] pred_named_entities = [] _, results_agg = compute_metrics(true_named_entities, pred_named_entities, ["PER"]) expected_agg = { "PER": { "strict": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "actual": 0, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "actual": 0, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "actual": 0, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "actual": 0, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, } } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] def test_compute_metrics_agg_scenario_2(): true_named_entities = [] pred_named_entities = [{"label": "PER", "start": 59, "end": 69}] _, results_agg = compute_metrics(true_named_entities, pred_named_entities, ["PER"]) expected_agg = { "PER": { "strict": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "actual": 1, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "actual": 1, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "actual": 1, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "actual": 1, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, } } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] def test_compute_metrics_agg_scenario_5(): true_named_entities = [{"label": "PER", "start": 59, "end": 69}] pred_named_entities = [{"label": "PER", "start": 57, "end": 69}] _, results_agg = compute_metrics(true_named_entities, pred_named_entities, ["PER"]) expected_agg = { "PER": { "strict": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 1, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, } } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] def test_compute_metrics_agg_scenario_4(): true_named_entities = [{"label": "PER", "start": 59, "end": 69}] pred_named_entities = [{"label": "LOC", "start": 59, "end": 69}] _, results_agg = compute_metrics(true_named_entities, pred_named_entities, ["PER", "LOC"]) expected_agg = { "PER": { "strict": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, }, "LOC": { "strict": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, }, } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] assert results_agg["LOC"] == expected_agg["LOC"] def test_compute_metrics_agg_scenario_1(): true_named_entities = [{"label": "PER", "start": 59, "end": 69}] pred_named_entities = [{"label": "PER", "start": 59, "end": 69}] _, results_agg = compute_metrics(true_named_entities, pred_named_entities, ["PER"]) expected_agg = { "PER": { "strict": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 1, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, } } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] def test_compute_metrics_agg_scenario_6(): true_named_entities = [{"label": "PER", "start": 59, "end": 69}] pred_named_entities = [{"label": "LOC", "start": 54, "end": 69}] _, results_agg = compute_metrics(true_named_entities, pred_named_entities, ["PER", "LOC"]) expected_agg = { "PER": { "strict": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 1, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 0, "actual": 1, "possible": 1, "precision": 0, "recall": 0, "f1": 0, }, }, "LOC": { "strict": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 0, "actual": 0, "possible": 0, "precision": 0, "recall": 0, "f1": 0, }, }, } assert results_agg["PER"]["strict"] == expected_agg["PER"]["strict"] assert results_agg["PER"]["ent_type"] == expected_agg["PER"]["ent_type"] assert results_agg["PER"]["partial"] == expected_agg["PER"]["partial"] assert results_agg["PER"]["exact"] == expected_agg["PER"]["exact"] assert results_agg["LOC"] == expected_agg["LOC"] def test_compute_metrics_extra_tags_in_prediction(): true_named_entities = [ {"label": "PER", "start": 50, "end": 52}, {"label": "ORG", "start": 59, "end": 69}, {"label": "ORG", "start": 71, "end": 72}, ] pred_named_entities = [ {"label": "LOC", "start": 50, "end": 52}, # Wrong type {"label": "ORG", "start": 59, "end": 69}, # Correct {"label": "MISC", "start": 71, "end": 72}, # Wrong type ] results, _ = compute_metrics(true_named_entities, pred_named_entities, ["PER", "LOC", "ORG"]) expected = { "strict": { "correct": 1, "incorrect": 1, "partial": 0, "missed": 1, "spurious": 0, "actual": 2, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 1, "incorrect": 1, "partial": 0, "missed": 1, "spurious": 0, "actual": 2, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "actual": 2, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 1, "spurious": 0, "actual": 2, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_compute_metrics_extra_tags_in_true(): true_named_entities = [ {"label": "PER", "start": 50, "end": 52}, {"label": "ORG", "start": 59, "end": 69}, {"label": "MISC", "start": 71, "end": 72}, ] pred_named_entities = [ {"label": "LOC", "start": 50, "end": 52}, # Wrong type {"label": "ORG", "start": 59, "end": 69}, # Correct {"label": "ORG", "start": 71, "end": 72}, # Spurious ] results, _ = compute_metrics(true_named_entities, pred_named_entities, ["PER", "LOC", "ORG"]) expected = { "strict": { "correct": 1, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 1, "actual": 3, "possible": 2, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 1, "incorrect": 1, "partial": 0, "missed": 0, "spurious": 1, "actual": 3, "possible": 2, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "actual": 3, "possible": 2, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 2, "incorrect": 0, "partial": 0, "missed": 0, "spurious": 1, "actual": 3, "possible": 2, "precision": 0, "recall": 0, "f1": 0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_compute_metrics_no_predictions(): true_named_entities = [ {"label": "PER", "start": 50, "end": 52}, {"label": "ORG", "start": 59, "end": 69}, {"label": "MISC", "start": 71, "end": 72}, ] pred_named_entities = [] results, _ = compute_metrics(true_named_entities, pred_named_entities, ["PER", "ORG", "MISC"]) expected = { "strict": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 3, "spurious": 0, "actual": 0, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, "ent_type": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 3, "spurious": 0, "actual": 0, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, "partial": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 3, "spurious": 0, "actual": 0, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, "exact": { "correct": 0, "incorrect": 0, "partial": 0, "missed": 3, "spurious": 0, "actual": 0, "possible": 3, "precision": 0, "recall": 0, "f1": 0, }, } assert results["strict"] == expected["strict"] assert results["ent_type"] == expected["ent_type"] assert results["partial"] == expected["partial"] assert results["exact"] == expected["exact"] def test_compute_actual_possible(): results = { "correct": 6, "incorrect": 3, "partial": 2, "missed": 4, "spurious": 2, } expected = { "correct": 6, "incorrect": 3, "partial": 2, "missed": 4, "spurious": 2, "possible": 15, "actual": 13, } out = compute_actual_possible(results) assert out == expected def test_compute_precision_recall(): results = { "correct": 6, "incorrect": 3, "partial": 2, "missed": 4, "spurious": 2, "possible": 15, "actual": 13, } expected = { "correct": 6, "incorrect": 3, "partial": 2, "missed": 4, "spurious": 2, "possible": 15, "actual": 13, "precision": 0.46153846153846156, "recall": 0.4, "f1": 0.42857142857142855, } out = compute_precision_recall(results) assert out == expected
23,628
27.851038
98
py
nervaluate
nervaluate-main/tests/test_utils.py
from nervaluate import ( collect_named_entities, conll_to_spans, find_overlap, list_to_spans, split_list, ) def test_list_to_spans(): before = [ ["O", "B-LOC", "I-LOC", "B-LOC", "I-LOC", "O"], ["O", "B-GPE", "I-GPE", "B-GPE", "I-GPE", "O"], ] expected = [ [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ], [ {"label": "GPE", "start": 1, "end": 2}, {"label": "GPE", "start": 3, "end": 4}, ], ] result = list_to_spans(before) assert result == expected def test_list_to_spans_1(): before = [ ["O", "O", "O", "O", "O", "O"], ["O", "O", "B-ORG", "I-ORG", "O", "O"], ["O", "O", "B-MISC", "I-MISC", "O", "O"], ] expected = [ [], [{"label": "ORG", "start": 2, "end": 3}], [{"label": "MISC", "start": 2, "end": 3}], ] actual = list_to_spans(before) assert actual == expected def test_conll_to_spans(): before = ( ",\tO\n" "Davos\tB-PER\n" "2018\tO\n" ":\tO\n" "Soros\tB-PER\n" "accuses\tO\n" "Trump\tB-PER\n" "of\tO\n" "wanting\tO\n" "\n" "foo\tO\n" ) after = [ [ {"label": "PER", "start": 1, "end": 1}, {"label": "PER", "start": 4, "end": 4}, {"label": "PER", "start": 6, "end": 6}, ], [], ] out = conll_to_spans(before) assert after == out def test_conll_to_spans_1(): before = ( "word\tO\nword\tO\nword\tO\nword\tO\nword\tO\nword\tO\n\n" "word\tO\nword\tO\nword\tB-ORG\nword\tI-ORG\nword\tO\nword\tO\n\n" "word\tO\nword\tO\nword\tB-MISC\nword\tI-MISC\nword\tO\nword\tO\n" ) expected = [ [], [{"label": "ORG", "start": 2, "end": 3}], [{"label": "MISC", "start": 2, "end": 3}], ] actual = conll_to_spans(before) assert actual == expected def test_split_list(): before = ["aa", "bb", "cc", "", "dd", "ee", "ff"] expected = [["aa", "bb", "cc"], ["dd", "ee", "ff"]] out = split_list(before) assert expected == out def test_collect_named_entities_same_type_in_sequence(): tags = ["O", "B-LOC", "I-LOC", "B-LOC", "I-LOC", "O"] result = collect_named_entities(tags) expected = [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ] assert result == expected def test_collect_named_entities_sequence_has_only_one_entity(): tags = ["B-LOC", "I-LOC"] result = collect_named_entities(tags) expected = [{"label": "LOC", "start": 0, "end": 1}] assert result == expected def test_collect_named_entities_entity_goes_until_last_token(): tags = ["O", "B-LOC", "I-LOC", "B-LOC", "I-LOC"] result = collect_named_entities(tags) expected = [ {"label": "LOC", "start": 1, "end": 2}, {"label": "LOC", "start": 3, "end": 4}, ] assert result == expected def test_collect_named_entities_no_entity(): tags = ["O", "O", "O", "O", "O"] result = collect_named_entities(tags) expected = [] assert result == expected def test_find_overlap_no_overlap(): pred_entity = {"label": "LOC", "start": 1, "end": 10} true_entity = {"label": "LOC", "start": 11, "end": 20} pred_range = range(pred_entity["start"], pred_entity["end"]) true_range = range(true_entity["start"], true_entity["end"]) pred_set = set(pred_range) true_set = set(true_range) intersect = find_overlap(pred_set, true_set) assert not intersect def test_find_overlap_total_overlap(): pred_entity = {"label": "LOC", "start": 10, "end": 22} true_entity = {"label": "LOC", "start": 11, "end": 20} pred_range = range(pred_entity["start"], pred_entity["end"]) true_range = range(true_entity["start"], true_entity["end"]) pred_set = set(pred_range) true_set = set(true_range) intersect = find_overlap(pred_set, true_set) assert intersect def test_find_overlap_start_overlap(): pred_entity = {"label": "LOC", "start": 5, "end": 12} true_entity = {"label": "LOC", "start": 11, "end": 20} pred_range = range(pred_entity["start"], pred_entity["end"]) true_range = range(true_entity["start"], true_entity["end"]) pred_set = set(pred_range) true_set = set(true_range) intersect = find_overlap(pred_set, true_set) assert intersect def test_find_overlap_end_overlap(): pred_entity = {"label": "LOC", "start": 15, "end": 25} true_entity = {"label": "LOC", "start": 11, "end": 20} pred_range = range(pred_entity["start"], pred_entity["end"]) true_range = range(true_entity["start"], true_entity["end"]) pred_set = set(pred_range) true_set = set(true_range) intersect = find_overlap(pred_set, true_set) assert intersect
4,946
24.111675
74
py
STTS
STTS-main/MViT/setup.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. from setuptools import find_packages, setup setup( name="slowfast", version="1.0", author="FAIR", url="unknown", description="SlowFast Video Understanding", install_requires=[ "yacs>=0.1.6", "pyyaml>=5.1", "av", "matplotlib", "termcolor>=1.1", "simplejson", "tqdm", "psutil", "matplotlib", "detectron2", "opencv-python", "pandas", "torchvision>=0.4.2", "pillow", "sklearn", "tensorboard", "fairscale", ], extras_require={"tensorboard_video_visualization": ["moviepy"]}, packages=find_packages(exclude=("configs", "tests")), )
794
23.090909
71
py
STTS
STTS-main/MViT/tools/run_net.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Wrapper to train and test a video classification model.""" from slowfast.config.defaults import assert_and_infer_cfg from slowfast.utils.misc import launch_job from slowfast.utils.parser import load_config, parse_args from demo_net import demo from test_net import test from train_net import train from visualization import visualize def get_func(cfg): train_func = train test_func = test return train_func, test_func def main(): """ Main function to spawn the train and test process. """ args = parse_args() cfg = load_config(args) cfg = assert_and_infer_cfg(cfg) # Perform training. if cfg.TRAIN.ENABLE: launch_job(cfg=cfg, init_method=args.init_method, func=train) # Perform multi-clip testing. if cfg.TEST.ENABLE: launch_job(cfg=cfg, init_method=args.init_method, func=test) # Perform model visualization. if cfg.TENSORBOARD.ENABLE and ( cfg.TENSORBOARD.MODEL_VIS.ENABLE or cfg.TENSORBOARD.WRONG_PRED_VIS.ENABLE ): launch_job(cfg=cfg, init_method=args.init_method, func=visualize) # Run demo. if cfg.DEMO.ENABLE: demo(cfg) if __name__ == "__main__": main()
1,303
24.568627
73
py
STTS
STTS-main/MViT/tools/visualization.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import numpy as np import pickle import torch import tqdm import slowfast.datasets.utils as data_utils import slowfast.utils.checkpoint as cu import slowfast.utils.distributed as du import slowfast.utils.logging as logging import slowfast.utils.misc as misc import slowfast.visualization.tensorboard_vis as tb from slowfast.datasets import loader from slowfast.models import build_model from slowfast.utils.env import pathmgr from slowfast.visualization.gradcam_utils import GradCAM from slowfast.visualization.prediction_vis import WrongPredictionVis from slowfast.visualization.utils import ( GetWeightAndActivation, process_layer_index_data, ) from slowfast.visualization.video_visualizer import VideoVisualizer logger = logging.get_logger(__name__) def run_visualization(vis_loader, model, cfg, writer=None): """ Run model visualization (weights, activations and model inputs) and visualize them on Tensorboard. Args: vis_loader (loader): video visualization loader. model (model): the video model to visualize. cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py writer (TensorboardWriter, optional): TensorboardWriter object to writer Tensorboard log. """ n_devices = cfg.NUM_GPUS * cfg.NUM_SHARDS prefix = "module/" if n_devices > 1 else "" # Get a list of selected layer names and indexing. layer_ls, indexing_dict = process_layer_index_data( cfg.TENSORBOARD.MODEL_VIS.LAYER_LIST, layer_name_prefix=prefix ) logger.info("Start Model Visualization.") # Register hooks for activations. model_vis = GetWeightAndActivation(model, layer_ls) if writer is not None and cfg.TENSORBOARD.MODEL_VIS.MODEL_WEIGHTS: layer_weights = model_vis.get_weights() writer.plot_weights_and_activations( layer_weights, tag="Layer Weights/", heat_map=False ) video_vis = VideoVisualizer( cfg.MODEL.NUM_CLASSES, cfg.TENSORBOARD.CLASS_NAMES_PATH, cfg.TENSORBOARD.MODEL_VIS.TOPK_PREDS, cfg.TENSORBOARD.MODEL_VIS.COLORMAP, ) if n_devices > 1: grad_cam_layer_ls = [ "module/" + layer for layer in cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST ] else: grad_cam_layer_ls = cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST if cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.ENABLE: gradcam = GradCAM( model, target_layers=grad_cam_layer_ls, data_mean=cfg.DATA.MEAN, data_std=cfg.DATA.STD, colormap=cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.COLORMAP, ) logger.info("Finish drawing weights.") global_idx = -1 for inputs, labels, _, meta in tqdm.tqdm(vis_loader): if cfg.NUM_GPUS: # Transfer the data to the current GPU device. if isinstance(inputs, (list,)): for i in range(len(inputs)): inputs[i] = inputs[i].cuda(non_blocking=True) else: inputs = inputs.cuda(non_blocking=True) labels = labels.cuda() for key, val in meta.items(): if isinstance(val, (list,)): for i in range(len(val)): val[i] = val[i].cuda(non_blocking=True) else: meta[key] = val.cuda(non_blocking=True) if cfg.DETECTION.ENABLE: activations, preds = model_vis.get_activations( inputs, meta["boxes"] ) else: activations, preds = model_vis.get_activations(inputs) if cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.ENABLE: if cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.USE_TRUE_LABEL: inputs, preds = gradcam(inputs, labels=labels) else: inputs, preds = gradcam(inputs) if cfg.NUM_GPUS: inputs = du.all_gather_unaligned(inputs) activations = du.all_gather_unaligned(activations) preds = du.all_gather_unaligned(preds) if isinstance(inputs[0], list): for i in range(len(inputs)): for j in range(len(inputs[0])): inputs[i][j] = inputs[i][j].cpu() else: inputs = [inp.cpu() for inp in inputs] preds = [pred.cpu() for pred in preds] else: inputs, activations, preds = [inputs], [activations], [preds] boxes = [None] * max(n_devices, 1) if cfg.DETECTION.ENABLE and cfg.NUM_GPUS: boxes = du.all_gather_unaligned(meta["boxes"]) boxes = [box.cpu() for box in boxes] if writer is not None: total_vids = 0 for i in range(max(n_devices, 1)): cur_input = inputs[i] cur_activations = activations[i] cur_batch_size = cur_input[0].shape[0] cur_preds = preds[i] cur_boxes = boxes[i] for cur_batch_idx in range(cur_batch_size): global_idx += 1 total_vids += 1 if ( cfg.TENSORBOARD.MODEL_VIS.INPUT_VIDEO or cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.ENABLE ): for path_idx, input_pathway in enumerate(cur_input): if cfg.TEST.DATASET == "ava" and cfg.AVA.BGR: video = input_pathway[ cur_batch_idx, [2, 1, 0], ... ] else: video = input_pathway[cur_batch_idx] if not cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.ENABLE: # Permute to (T, H, W, C) from (C, T, H, W). video = video.permute(1, 2, 3, 0) video = data_utils.revert_tensor_normalize( video, cfg.DATA.MEAN, cfg.DATA.STD ) else: # Permute from (T, C, H, W) to (T, H, W, C) video = video.permute(0, 2, 3, 1) bboxes = ( None if cur_boxes is None else cur_boxes[:, 1:] ) cur_prediction = ( cur_preds if cfg.DETECTION.ENABLE else cur_preds[cur_batch_idx] ) video = video_vis.draw_clip( video, cur_prediction, bboxes=bboxes ) video = ( torch.from_numpy(np.array(video)) .permute(0, 3, 1, 2) .unsqueeze(0) ) writer.add_video( video, tag="Input {}/Pathway {}".format( global_idx, path_idx + 1 ), ) if cfg.TENSORBOARD.MODEL_VIS.ACTIVATIONS: writer.plot_weights_and_activations( cur_activations, tag="Input {}/Activations: ".format(global_idx), batch_idx=cur_batch_idx, indexing_dict=indexing_dict, ) def perform_wrong_prediction_vis(vis_loader, model, cfg): """ Visualize video inputs with wrong predictions on Tensorboard. Args: vis_loader (loader): video visualization loader. model (model): the video model to visualize. cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py """ wrong_prediction_visualizer = WrongPredictionVis(cfg=cfg) for batch_idx, (inputs, labels, _, _) in tqdm.tqdm(enumerate(vis_loader)): if cfg.NUM_GPUS: # Transfer the data to the current GPU device. if isinstance(inputs, (list,)): for i in range(len(inputs)): inputs[i] = inputs[i].cuda(non_blocking=True) else: inputs = inputs.cuda(non_blocking=True) labels = labels.cuda() # Some model modify the original input. inputs_clone = [inp.clone() for inp in inputs] preds = model(inputs) if cfg.NUM_GPUS > 1: preds, labels = du.all_gather([preds, labels]) if isinstance(inputs_clone, (list,)): inputs_clone = du.all_gather(inputs_clone) else: inputs_clone = du.all_gather([inputs_clone])[0] if cfg.NUM_GPUS: # Transfer the data to the current CPU device. labels = labels.cpu() preds = preds.cpu() if isinstance(inputs_clone, (list,)): for i in range(len(inputs_clone)): inputs_clone[i] = inputs_clone[i].cpu() else: inputs_clone = inputs_clone.cpu() # If using CPU (NUM_GPUS = 0), 1 represent 1 CPU. n_devices = max(cfg.NUM_GPUS, 1) for device_idx in range(1, n_devices + 1): wrong_prediction_visualizer.visualize_vid( video_input=inputs_clone, labels=labels, preds=preds.detach().clone(), batch_idx=device_idx * batch_idx, ) logger.info( "Class indices with wrong predictions: {}".format( sorted(wrong_prediction_visualizer.wrong_class_prediction) ) ) wrong_prediction_visualizer.clean() def visualize(cfg): """ Perform layer weights and activations visualization on the model. Args: cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py """ if cfg.TENSORBOARD.ENABLE and ( cfg.TENSORBOARD.MODEL_VIS.ENABLE or cfg.TENSORBOARD.WRONG_PRED_VIS.ENABLE ): # Set up environment. du.init_distributed_training(cfg) # Set random seed from configs. np.random.seed(cfg.RNG_SEED) torch.manual_seed(cfg.RNG_SEED) # Setup logging format. logging.setup_logging(cfg.OUTPUT_DIR) # Print config. logger.info("Model Visualization with config:") logger.info(cfg) # Build the video model and print model statistics. model = build_model(cfg) model.eval() if du.is_master_proc() and cfg.LOG_MODEL_INFO: misc.log_model_info(model, cfg, use_train_input=False) cu.load_test_checkpoint(cfg, model) # Create video testing loaders. vis_loader = loader.construct_loader(cfg, "test") if cfg.DETECTION.ENABLE: assert cfg.NUM_GPUS == cfg.TEST.BATCH_SIZE or cfg.NUM_GPUS == 0 # Set up writer for logging to Tensorboard format. if du.is_master_proc(cfg.NUM_GPUS * cfg.NUM_SHARDS): writer = tb.TensorboardWriter(cfg) else: writer = None if cfg.TENSORBOARD.PREDICTIONS_PATH != "": assert not cfg.DETECTION.ENABLE, "Detection is not supported." logger.info( "Visualizing class-level performance from saved results..." ) if writer is not None: with pathmgr.open(cfg.TENSORBOARD.PREDICTIONS_PATH, "rb") as f: preds, labels = pickle.load(f, encoding="latin1") writer.plot_eval(preds, labels) if cfg.TENSORBOARD.MODEL_VIS.ENABLE: if cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.ENABLE: assert ( not cfg.DETECTION.ENABLE ), "Detection task is currently not supported for Grad-CAM visualization." if cfg.MODEL.ARCH in cfg.MODEL.SINGLE_PATHWAY_ARCH: assert ( len(cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST) == 1 ), "The number of chosen CNN layers must be equal to the number of pathway(s), given {} layer(s).".format( len(cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST) ) elif cfg.MODEL.ARCH in cfg.MODEL.MULTI_PATHWAY_ARCH: assert ( len(cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST) == 2 ), "The number of chosen CNN layers must be equal to the number of pathway(s), given {} layer(s).".format( len(cfg.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST) ) else: raise NotImplementedError( "Model arch {} is not in {}".format( cfg.MODEL.ARCH, cfg.MODEL.SINGLE_PATHWAY_ARCH + cfg.MODEL.MULTI_PATHWAY_ARCH, ) ) logger.info( "Visualize model analysis for {} iterations".format( len(vis_loader) ) ) # Run visualization on the model run_visualization(vis_loader, model, cfg, writer) if cfg.TENSORBOARD.WRONG_PRED_VIS.ENABLE: logger.info( "Visualize Wrong Predictions for {} iterations".format( len(vis_loader) ) ) perform_wrong_prediction_vis(vis_loader, model, cfg) if writer is not None: writer.close()
14,000
39.465318
126
py
STTS
STTS-main/MViT/tools/benchmark.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved """ A script to benchmark data loading. """ import slowfast.utils.logging as logging from slowfast.utils.benchmark import benchmark_data_loading from slowfast.utils.misc import launch_job from slowfast.utils.parser import load_config, parse_args logger = logging.get_logger(__name__) def main(): args = parse_args() cfg = load_config(args) launch_job( cfg=cfg, init_method=args.init_method, func=benchmark_data_loading ) if __name__ == "__main__": main()
584
21.5
74
py
STTS
STTS-main/MViT/tools/test_net.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Multi-view test a video classification model.""" import numpy as np import os import pickle import torch import torch.nn.functional as F from einops import rearrange import slowfast.utils.checkpoint as cu import slowfast.utils.distributed as du import slowfast.utils.logging as logging import slowfast.utils.misc as misc import slowfast.visualization.tensorboard_vis as tb from slowfast.datasets import loader from slowfast.models import build_model from slowfast.utils.env import pathmgr from slowfast.utils.meters import AVAMeter, TestMeter logger = logging.get_logger(__name__) @torch.no_grad() def perform_test(test_loader, model, test_meter, cfg, writer=None): """ For classification: Perform mutli-view testing that uniformly samples N clips from a video along its temporal axis. For each clip, it takes 3 crops to cover the spatial dimension, followed by averaging the softmax scores across all Nx3 views to form a video-level prediction. All video predictions are compared to ground-truth labels and the final testing performance is logged. For detection: Perform fully-convolutional testing on the full frames without crop. Args: test_loader (loader): video testing loader. model (model): the pretrained video model to test. test_meter (TestMeter): testing meters to log and ensemble the testing results. cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py writer (TensorboardWriter object, optional): TensorboardWriter object to writer Tensorboard log. """ # Enable eval mode. model.eval() test_meter.iter_tic() for cur_iter, (inputs, labels, video_idx, meta) in enumerate(test_loader): if cfg.NUM_GPUS: # Transfer the data to the current GPU device. if isinstance(inputs, (list,)): for i in range(len(inputs)): inputs[i] = inputs[i].cuda(non_blocking=True) else: inputs = inputs.cuda(non_blocking=True) # Transfer the data to the current GPU device. labels = labels.cuda() video_idx = video_idx.cuda() for key, val in meta.items(): if isinstance(val, (list,)): for i in range(len(val)): val[i] = val[i].cuda(non_blocking=True) else: meta[key] = val.cuda(non_blocking=True) test_meter.data_toc() if cfg.DETECTION.ENABLE: # Compute the predictions. preds = model(inputs, meta["boxes"]) ori_boxes = meta["ori_boxes"] metadata = meta["metadata"] preds = preds.detach().cpu() if cfg.NUM_GPUS else preds.detach() ori_boxes = ( ori_boxes.detach().cpu() if cfg.NUM_GPUS else ori_boxes.detach() ) metadata = ( metadata.detach().cpu() if cfg.NUM_GPUS else metadata.detach() ) if cfg.NUM_GPUS > 1: preds = torch.cat(du.all_gather_unaligned(preds), dim=0) ori_boxes = torch.cat(du.all_gather_unaligned(ori_boxes), dim=0) metadata = torch.cat(du.all_gather_unaligned(metadata), dim=0) test_meter.iter_toc() # Update and log stats. test_meter.update_stats(preds, ori_boxes, metadata) test_meter.log_iter_stats(None, cur_iter) else: # Perform the forward pass. preds = model(inputs) # Gather all the predictions across all the devices to perform ensemble. if cfg.NUM_GPUS > 1: preds, labels, video_idx = du.all_gather( [preds, labels, video_idx] ) if cfg.NUM_GPUS: preds = preds.cpu() labels = labels.cpu() video_idx = video_idx.cpu() test_meter.iter_toc() # Update and log stats. test_meter.update_stats( preds.detach(), labels.detach(), video_idx.detach() ) test_meter.log_iter_stats(cur_iter) test_meter.iter_tic() # Log epoch stats and print the final testing results. if not cfg.DETECTION.ENABLE: all_preds = test_meter.video_preds.clone().detach() all_labels = test_meter.video_labels if cfg.NUM_GPUS: all_preds = all_preds.cpu() all_labels = all_labels.cpu() if writer is not None: writer.plot_eval(preds=all_preds, labels=all_labels) if cfg.TEST.SAVE_RESULTS_PATH != "": save_path = os.path.join(cfg.OUTPUT_DIR, cfg.TEST.SAVE_RESULTS_PATH) if du.is_root_proc(): with pathmgr.open(save_path, "wb") as f: pickle.dump([all_preds, all_labels], f) logger.info( "Successfully saved prediction results to {}".format(save_path) ) test_meter.finalize_metrics() return test_meter def test(cfg): """ Perform multi-view testing on the pretrained video model. Args: cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py """ # Set up environment. du.init_distributed_training(cfg) # Set random seed from configs. np.random.seed(cfg.RNG_SEED) torch.manual_seed(cfg.RNG_SEED) # Setup logging format. logging.setup_logging(cfg.OUTPUT_DIR) # Build the video model and print model statistics. model = build_model(cfg) if du.is_master_proc() and cfg.LOG_MODEL_INFO: misc.log_model_info(model, cfg, use_train_input=False) cu.load_test_checkpoint(cfg, model) # Create video testing loaders. test_loader = loader.construct_loader(cfg, "test") logger.info("Testing model for {} iterations".format(len(test_loader))) if cfg.DETECTION.ENABLE: assert cfg.NUM_GPUS == cfg.TEST.BATCH_SIZE or cfg.NUM_GPUS == 0 test_meter = AVAMeter(len(test_loader), cfg, mode="test") else: assert ( test_loader.dataset.num_videos % (cfg.TEST.NUM_ENSEMBLE_VIEWS * cfg.TEST.NUM_SPATIAL_CROPS) == 0 ) # Create meters for multi-view testing. test_meter = TestMeter( test_loader.dataset.num_videos // (cfg.TEST.NUM_ENSEMBLE_VIEWS * cfg.TEST.NUM_SPATIAL_CROPS), cfg.TEST.NUM_ENSEMBLE_VIEWS * cfg.TEST.NUM_SPATIAL_CROPS, cfg.MODEL.NUM_CLASSES, len(test_loader), cfg.DATA.MULTI_LABEL, cfg.DATA.ENSEMBLE_METHOD, ) # Set up writer for logging to Tensorboard format. if cfg.TENSORBOARD.ENABLE and du.is_master_proc( cfg.NUM_GPUS * cfg.NUM_SHARDS ): writer = tb.TensorboardWriter(cfg) else: writer = None # # Perform multi-view test on the entire dataset. test_meter = perform_test(test_loader, model, test_meter, cfg, writer) if writer is not None: writer.close()
7,234
34.816832
84
py
STTS
STTS-main/MViT/tools/demo_net.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import numpy as np import time import torch import tqdm from slowfast.utils import logging from slowfast.visualization.async_predictor import AsyncDemo, AsyncVis from slowfast.visualization.ava_demo_precomputed_boxes import ( AVAVisualizerWithPrecomputedBox, ) from slowfast.visualization.demo_loader import ThreadVideoManager, VideoManager from slowfast.visualization.predictor import ActionPredictor from slowfast.visualization.video_visualizer import VideoVisualizer logger = logging.get_logger(__name__) def run_demo(cfg, frame_provider): """ Run demo visualization. Args: cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py frame_provider (iterator): Python iterator that return task objects that are filled with necessary information such as `frames`, `id` and `num_buffer_frames` for the prediction and visualization pipeline. """ # Set random seed from configs. np.random.seed(cfg.RNG_SEED) torch.manual_seed(cfg.RNG_SEED) # Setup logging format. logging.setup_logging(cfg.OUTPUT_DIR) # Print config. logger.info("Run demo with config:") logger.info(cfg) common_classes = ( cfg.DEMO.COMMON_CLASS_NAMES if len(cfg.DEMO.LABEL_FILE_PATH) != 0 else None ) video_vis = VideoVisualizer( num_classes=cfg.MODEL.NUM_CLASSES, class_names_path=cfg.DEMO.LABEL_FILE_PATH, top_k=cfg.TENSORBOARD.MODEL_VIS.TOPK_PREDS, thres=cfg.DEMO.COMMON_CLASS_THRES, lower_thres=cfg.DEMO.UNCOMMON_CLASS_THRES, common_class_names=common_classes, colormap=cfg.TENSORBOARD.MODEL_VIS.COLORMAP, mode=cfg.DEMO.VIS_MODE, ) async_vis = AsyncVis(video_vis, n_workers=cfg.DEMO.NUM_VIS_INSTANCES) if cfg.NUM_GPUS <= 1: model = ActionPredictor(cfg=cfg, async_vis=async_vis) else: model = AsyncDemo(cfg=cfg, async_vis=async_vis) seq_len = cfg.DATA.NUM_FRAMES * cfg.DATA.SAMPLING_RATE assert ( cfg.DEMO.BUFFER_SIZE <= seq_len // 2 ), "Buffer size cannot be greater than half of sequence length." num_task = 0 # Start reading frames. frame_provider.start() for able_to_read, task in frame_provider: if not able_to_read: break if task is None: time.sleep(0.02) continue num_task += 1 model.put(task) try: task = model.get() num_task -= 1 yield task except IndexError: continue while num_task != 0: try: task = model.get() num_task -= 1 yield task except IndexError: continue def demo(cfg): """ Run inference on an input video or stream from webcam. Args: cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py """ # AVA format-specific visualization with precomputed boxes. if cfg.DETECTION.ENABLE and cfg.DEMO.PREDS_BOXES != "": precomputed_box_vis = AVAVisualizerWithPrecomputedBox(cfg) precomputed_box_vis() else: start = time.time() if cfg.DEMO.THREAD_ENABLE: frame_provider = ThreadVideoManager(cfg) else: frame_provider = VideoManager(cfg) for task in tqdm.tqdm(run_demo(cfg, frame_provider)): frame_provider.display(task) frame_provider.join() frame_provider.clean() logger.info("Finish demo in: {}".format(time.time() - start))
3,683
29.7
93
py
STTS
STTS-main/MViT/tools/submit.py
import argparse import os from pathlib import Path import submitit import torch from slowfast.utils.misc import launch_job from slowfast.utils.parser import load_config from run_net import get_func def parse_args(): parser = argparse.ArgumentParser( "Submitit for onestage training", add_help=False ) parser.add_argument( "--num_gpus", help="Number of GPUs", default=8, type=int, ) parser.add_argument( "--num_shards", help="Number of Nodes", default=1, type=int, ) parser.add_argument( "--partition", default="learnfair", type=str, help="Partition where to submit" ) parser.add_argument("--timeout", default=60 * 72, type=int, help="Duration of the job") parser.add_argument("--cfg", dest="cfg_file", help="Path to the config file", default="configs/test_R50_8GPU.yaml", type=str) parser.add_argument( "--job_dir", default="", type=str, help="Job dir. Leave empty for automatic." ) parser.add_argument( "--name", default="", type=str, help="Job dir. Leave empty for automatic." ) parser.add_argument( "--resume-from", default="", type=str, help=( "Weights to resume from (.*pth file) or a file (last_checkpoint) that contains " + "weight file name from the same directory" ), ) parser.add_argument("--resume-job", default="", type=str, help="resume training from the job") parser.add_argument("--use_volta32", action='store_true', help="Big models? Use this") parser.add_argument("--postfix", default="experiment", type=str, help="Postfix of the jobs") parser.add_argument("--mail", default="", type=str, help="Email this user when the job finishes if specified") parser.add_argument('--comment', default="", type=str, help='Comment to pass to scheduler, e.g. priority message') parser.add_argument( "opts", help="See lib/config/defaults.py for all options", default=None, nargs=argparse.REMAINDER, ) return parser.parse_args() def get_shared_folder() -> Path: user = os.getenv("USER") if Path("/checkpoint/").is_dir(): p = Path(f"/checkpoint/{user}/experiments/TopkMVIT") p.mkdir(exist_ok=True) return p raise RuntimeError("No shared folder available") def launch(shard_id, num_shards, cfg, init_method): os.environ["NCCL_MIN_NRINGS"] = "8" print ("Pytorch version: ", torch.__version__) cfg.SHARD_ID = shard_id cfg.NUM_SHARDS = num_shards print([ shard_id, num_shards, cfg ]) train, test = get_func(cfg) # Launch job. if cfg.TRAIN.ENABLE: launch_job(cfg=cfg, init_method=init_method, func=train) if cfg.TEST.ENABLE: launch_job(cfg=cfg, init_method=init_method, func=test) class Trainer(object): def __init__(self, args): self.args = args def __call__(self): socket_name = os.popen("ip r | grep default | awk '{print $5}'").read().strip('\n') print("Setting GLOO and NCCL sockets IFNAME to: {}".format(socket_name)) os.environ["GLOO_SOCKET_IFNAME"] = socket_name # not sure if the next line is really affect anything os.environ["NCCL_SOCKET_IFNAME"] = socket_name hostname_first_node = os.popen( "scontrol show hostnames $SLURM_JOB_NODELIST" ).read().split("\n")[0] dist_url = "tcp://{}:12399".format(hostname_first_node) print("We will use the following dist url: {}".format(dist_url)) self._setup_gpu_args() results = launch( shard_id=self.args.machine_rank, num_shards=self.args.num_shards, cfg=load_config(self.args), init_method=dist_url, ) return results def checkpoint(self): import submitit job_env = submitit.JobEnvironment() slurm_job_id = job_env.job_id if self.args.resume_job == "": self.args.resume_job = slurm_job_id print("Requeuing ", self.args) empty_trainer = type(self)(self.args) return submitit.helpers.DelayedSubmission(empty_trainer) def _setup_gpu_args(self): import submitit job_env = submitit.JobEnvironment() self.args.output_dir = str(Path(str(self.args.output_dir).replace("%j", str(job_env.job_id)))) print(self.args) self.args.machine_rank = job_env.global_rank print(f"Process rank: {job_env.global_rank}") def main(): args = parse_args() if args.name == "": cfg_name = os.path.splitext(os.path.basename(args.cfg_file))[0] args.name = '_'.join([cfg_name, args.postfix]) if args.job_dir == "": args.job_dir = get_shared_folder() / "%j" # Note that the folder will depend on the job_id, to easily track experiments #executor = submitit.AutoExecutor(folder=Path(args.job_dir) / "%j", slurm_max_num_timeout=30) executor = submitit.AutoExecutor(folder=args.job_dir, slurm_max_num_timeout=30) # cluster setup is defined by environment variables num_gpus_per_node = args.num_gpus nodes = args.num_shards partition = args.partition timeout_min = args.timeout kwargs = {} if args.use_volta32: kwargs['slurm_constraint'] = 'volta32gb,ib4' if args.comment: kwargs['slurm_comment'] = args.comment executor.update_parameters( mem_gb=60 * num_gpus_per_node, gpus_per_node=num_gpus_per_node, tasks_per_node=1, cpus_per_task=10 * num_gpus_per_node, nodes=nodes, timeout_min=timeout_min, # max is 60 * 72 slurm_partition=partition, slurm_signal_delay_s=120, **kwargs ) print(args.name) executor.update_parameters(name=args.name) args.output_dir = args.job_dir trainer = Trainer(args) job = executor.submit(trainer) print("Submitted job_id:", job.job_id) if __name__ == "__main__": main()
6,114
30.040609
102
py
STTS
STTS-main/MViT/tools/train_net.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Train a video classification model.""" import numpy as np import torch from fvcore.nn.precise_bn import get_bn_modules, update_bn_stats import copy import slowfast.models.losses as losses import slowfast.models.optimizer as optim import slowfast.utils.checkpoint as cu import slowfast.utils.distributed as du import slowfast.utils.logging as logging import slowfast.utils.metrics as metrics import slowfast.utils.misc as misc import slowfast.visualization.tensorboard_vis as tb from slowfast.datasets import loader from slowfast.datasets.mixup import MixUp from slowfast.models import build_model from slowfast.utils.meters import AVAMeter, EpochTimer, TrainMeter, ValMeter from slowfast.utils.multigrid import MultigridSchedule logger = logging.get_logger(__name__) def train_epoch( train_loader, model, loss_fun, optimizer, scaler, train_meter, cur_epoch, total_epochs, cfg, writer=None, ): """ Perform the video training for one epoch. Args: train_loader (loader): video training loader. model (model): the video model to train. optimizer (optim): the optimizer to perform optimization on the model's parameters. train_meter (TrainMeter): training meters to log the training performance. cur_epoch (int): current epoch of training. cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py writer (TensorboardWriter, optional): TensorboardWriter object to writer Tensorboard log. """ # Enable train mode. model.train() train_meter.iter_tic() data_size = len(train_loader) if hasattr(model, "module"): model_noddp = model.module else: model_noddp = model total_steps = data_size * total_epochs if cfg.MIXUP.ENABLE: mixup_fn = MixUp( mixup_alpha=cfg.MIXUP.ALPHA, cutmix_alpha=cfg.MIXUP.CUTMIX_ALPHA, mix_prob=cfg.MIXUP.PROB, switch_prob=cfg.MIXUP.SWITCH_PROB, label_smoothing=cfg.MIXUP.LABEL_SMOOTH_VALUE, num_classes=cfg.MODEL.NUM_CLASSES, ) for cur_iter, (inputs, labels, _, meta) in enumerate(train_loader): # Transfer the data to the current GPU device. if cfg.NUM_GPUS: if isinstance(inputs, (list,)): for i in range(len(inputs)): inputs[i] = inputs[i].cuda(non_blocking=True) else: inputs = inputs.cuda(non_blocking=True) labels = labels.cuda() for key, val in meta.items(): if isinstance(val, (list,)): for i in range(len(val)): val[i] = val[i].cuda(non_blocking=True) else: meta[key] = val.cuda(non_blocking=True) # Update the learning rate. lr = optim.get_epoch_lr(cur_epoch + float(cur_iter) / data_size, cfg) optim.set_lr(optimizer, lr, cfg) train_meter.data_toc() if cfg.MIXUP.ENABLE: samples, labels = mixup_fn(inputs[0], labels) inputs[0] = samples cur_step = cur_epoch * data_size + cur_iter if hasattr(model_noddp, 'update_sigma') and cfg.MVIT.DECAY_SIGMA: model_noddp.update_sigma(cur_step, total_steps) with torch.cuda.amp.autocast(enabled=cfg.TRAIN.MIXED_PRECISION): if cfg.DETECTION.ENABLE: preds = model(inputs, meta["boxes"]) else: preds = model(inputs) # Compute the loss. loss = loss_fun(preds, labels) # check Nan Loss. misc.check_nan_losses(loss) # Perform the backward pass. optimizer.zero_grad() scaler.scale(loss).backward() # Unscales the gradients of optimizer's assigned params in-place scaler.unscale_(optimizer) # Clip gradients if necessary if cfg.SOLVER.CLIP_GRAD_VAL: torch.nn.utils.clip_grad_value_( model.parameters(), cfg.SOLVER.CLIP_GRAD_VAL ) elif cfg.SOLVER.CLIP_GRAD_L2NORM: torch.nn.utils.clip_grad_norm_( model.parameters(), cfg.SOLVER.CLIP_GRAD_L2NORM ) # Update the parameters. scaler.step(optimizer) scaler.update() if isinstance(preds, (tuple,)): preds = preds[0] if cfg.MIXUP.ENABLE: _top_max_k_vals, top_max_k_inds = torch.topk( labels, 2, dim=1, largest=True, sorted=True ) idx_top1 = torch.arange(labels.shape[0]), top_max_k_inds[:, 0] idx_top2 = torch.arange(labels.shape[0]), top_max_k_inds[:, 1] preds = preds.detach() preds[idx_top1] += preds[idx_top2] preds[idx_top2] = 0.0 labels = top_max_k_inds[:, 0] if cfg.DETECTION.ENABLE: if cfg.NUM_GPUS > 1: loss = du.all_reduce([loss])[0] loss = loss.item() # Update and log stats. train_meter.update_stats(None, None, None, loss, lr) # write to tensorboard format if available. if writer is not None: writer.add_scalars( {"Train/loss": loss, "Train/lr": lr}, global_step=data_size * cur_epoch + cur_iter, ) else: top1_err, top5_err = None, None if cfg.DATA.MULTI_LABEL: # Gather all the predictions across all the devices. if cfg.NUM_GPUS > 1: [loss] = du.all_reduce([loss]) loss = loss.item() else: # Compute the errors. num_topks_correct = metrics.topks_correct(preds, labels, (1, 5)) top1_err, top5_err = [ (1.0 - x / preds.size(0)) * 100.0 for x in num_topks_correct ] # Gather all the predictions across all the devices. if cfg.NUM_GPUS > 1: loss, top1_err, top5_err = du.all_reduce( [loss, top1_err, top5_err] ) # Copy the stats from GPU to CPU (sync point). loss, top1_err, top5_err = ( loss.item(), top1_err.item(), top5_err.item(), ) # Update and log stats. train_meter.update_stats( top1_err, top5_err, loss, lr, inputs[0].size(0) * max( cfg.NUM_GPUS, 1 ), # If running on CPU (cfg.NUM_GPUS == 1), use 1 to represent 1 CPU. ) # write to tensorboard format if available. if writer is not None: writer.add_scalars( { "Train/loss": loss, "Train/lr": lr, "Train/Top1_err": top1_err, "Train/Top5_err": top5_err, }, global_step=data_size * cur_epoch + cur_iter, ) train_meter.iter_toc() # measure allreduce for this meter train_meter.log_iter_stats(cur_epoch, cur_iter) train_meter.iter_tic() # Log epoch stats. train_meter.log_epoch_stats(cur_epoch) train_meter.reset() @torch.no_grad() def eval_epoch(val_loader, model, val_meter, cur_epoch, cfg, writer=None): """ Evaluate the model on the val set. Args: val_loader (loader): data loader to provide validation data. model (model): model to evaluate the performance. val_meter (ValMeter): meter instance to record and calculate the metrics. cur_epoch (int): number of the current epoch of training. cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py writer (TensorboardWriter, optional): TensorboardWriter object to writer Tensorboard log. """ # Evaluation mode enabled. The running stats would not be updated. model.eval() val_meter.iter_tic() if hasattr(model, "module"): model_noddp = model.module else: model_noddp = model for cur_iter, (inputs, labels, _, meta) in enumerate(val_loader): if cfg.NUM_GPUS: # Transferthe data to the current GPU device. if isinstance(inputs, (list,)): for i in range(len(inputs)): inputs[i] = inputs[i].cuda(non_blocking=True) else: inputs = inputs.cuda(non_blocking=True) labels = labels.cuda() for key, val in meta.items(): if isinstance(val, (list,)): for i in range(len(val)): val[i] = val[i].cuda(non_blocking=True) else: meta[key] = val.cuda(non_blocking=True) val_meter.data_toc() if hasattr(model_noddp, 'set_warm'): if cfg.MVIT.USE_WARMUP and (cur_epoch < cfg.MVIT.SCORE_WARMUP_EPOCH): model_noddp.set_warm(flag=True) else: model_noddp.set_warm(flag=False) if hasattr(model_noddp, 'update_scale_rate') and cfg.MVIT.CURRICULUM: model_noddp.update_scale_rate(1.0) if cfg.DETECTION.ENABLE: # Compute the predictions. preds = model(inputs, meta["boxes"]) ori_boxes = meta["ori_boxes"] metadata = meta["metadata"] if cfg.NUM_GPUS: preds = preds.cpu() ori_boxes = ori_boxes.cpu() metadata = metadata.cpu() if cfg.NUM_GPUS > 1: preds = torch.cat(du.all_gather_unaligned(preds), dim=0) ori_boxes = torch.cat(du.all_gather_unaligned(ori_boxes), dim=0) metadata = torch.cat(du.all_gather_unaligned(metadata), dim=0) val_meter.iter_toc() # Update and log stats. val_meter.update_stats(preds, ori_boxes, metadata) else: preds = model(inputs) if cfg.DATA.MULTI_LABEL: if cfg.NUM_GPUS > 1: preds, labels = du.all_gather([preds, labels]) else: # Compute the errors. num_topks_correct = metrics.topks_correct(preds, labels, (1, 5)) # Combine the errors across the GPUs. top1_err, top5_err = [ (1.0 - x / preds.size(0)) * 100.0 for x in num_topks_correct ] if cfg.NUM_GPUS > 1: top1_err, top5_err = du.all_reduce([top1_err, top5_err]) # Copy the errors from GPU to CPU (sync point). top1_err, top5_err = top1_err.item(), top5_err.item() val_meter.iter_toc() # Update and log stats. val_meter.update_stats( top1_err, top5_err, inputs[0].size(0) * max( cfg.NUM_GPUS, 1 ), # If running on CPU (cfg.NUM_GPUS == 1), use 1 to represent 1 CPU. ) # write to tensorboard format if available. if writer is not None: writer.add_scalars( {"Val/Top1_err": top1_err, "Val/Top5_err": top5_err}, global_step=len(val_loader) * cur_epoch + cur_iter, ) val_meter.update_predictions(preds, labels) val_meter.log_iter_stats(cur_epoch, cur_iter) val_meter.iter_tic() # Log epoch stats. val_meter.log_epoch_stats(cur_epoch) # write to tensorboard format if available. if writer is not None: if cfg.DETECTION.ENABLE: writer.add_scalars( {"Val/mAP": val_meter.full_map}, global_step=cur_epoch ) else: all_preds = [pred.clone().detach() for pred in val_meter.all_preds] all_labels = [ label.clone().detach() for label in val_meter.all_labels ] if cfg.NUM_GPUS: all_preds = [pred.cpu() for pred in all_preds] all_labels = [label.cpu() for label in all_labels] writer.plot_eval( preds=all_preds, labels=all_labels, global_step=cur_epoch ) val_meter.reset() def calculate_and_update_precise_bn(loader, model, num_iters=200, use_gpu=True): """ Update the stats in bn layers by calculate the precise stats. Args: loader (loader): data loader to provide training data. model (model): model to update the bn stats. num_iters (int): number of iterations to compute and update the bn stats. use_gpu (bool): whether to use GPU or not. """ def _gen_loader(): for inputs, *_ in loader: if use_gpu: if isinstance(inputs, (list,)): for i in range(len(inputs)): inputs[i] = inputs[i].cuda(non_blocking=True) else: inputs = inputs.cuda(non_blocking=True) yield inputs # Update the bn stats. update_bn_stats(model, _gen_loader(), num_iters) def build_trainer(cfg): """ Build training model and its associated tools, including optimizer, dataloaders and meters. Args: cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py Returns: model (nn.Module): training model. optimizer (Optimizer): optimizer. train_loader (DataLoader): training data loader. val_loader (DataLoader): validatoin data loader. precise_bn_loader (DataLoader): training data loader for computing precise BN. train_meter (TrainMeter): tool for measuring training stats. val_meter (ValMeter): tool for measuring validation stats. """ # Build the video model and print model statistics. model = build_model(cfg) if du.is_master_proc() and cfg.LOG_MODEL_INFO: misc.log_model_info(model, cfg, use_train_input=True) # Construct the optimizer. optimizer = optim.construct_optimizer(model, cfg) # Create the video train and val loaders. train_loader = loader.construct_loader(cfg, "train") val_loader = loader.construct_loader(cfg, "val") precise_bn_loader = loader.construct_loader( cfg, "train", is_precise_bn=True ) # Create meters. train_meter = TrainMeter(len(train_loader), cfg) val_meter = ValMeter(len(val_loader), cfg) return ( model, optimizer, train_loader, val_loader, precise_bn_loader, train_meter, val_meter, ) def train(cfg): """ Train a video model for many epochs on train set and evaluate it on val set. Args: cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py """ # Set up environment. du.init_distributed_training(cfg) # Set random seed from configs. np.random.seed(cfg.RNG_SEED) torch.manual_seed(cfg.RNG_SEED) # Setup logging format. logging.setup_logging(cfg.OUTPUT_DIR) # Init multigrid. multigrid = None if cfg.MULTIGRID.LONG_CYCLE or cfg.MULTIGRID.SHORT_CYCLE: multigrid = MultigridSchedule() cfg = multigrid.init_multigrid(cfg) if cfg.MULTIGRID.LONG_CYCLE: cfg, _ = multigrid.update_long_cycle(cfg, cur_epoch=0) # Build the video model and print model statistics. model = build_model(cfg) if du.is_master_proc() and cfg.LOG_MODEL_INFO: misc.log_model_info(model, cfg, use_train_input=True) loss_fun = losses.get_loss_func(cfg.MODEL.LOSS_FUNC)( reduction="mean" ) # Construct the optimizer. optimizer = optim.construct_optimizer(model, cfg) # Create a GradScaler for mixed precision training scaler = torch.cuda.amp.GradScaler(enabled=cfg.TRAIN.MIXED_PRECISION) # Load a checkpoint to resume training if applicable. if not cfg.TRAIN.FINETUNE: start_epoch = cu.load_train_checkpoint(cfg, model, optimizer, scaler if cfg.TRAIN.MIXED_PRECISION else None) else: if cfg.TRAIN.AUTO_RESUME and cu.has_checkpoint(cfg.OUTPUT_DIR): last_checkpoint = cu.get_last_checkpoint(cfg.OUTPUT_DIR) checkpoint_epoch = cu.load_checkpoint(last_checkpoint, model, cfg.NUM_GPUS > 1, optimizer, scaler if cfg.TRAIN.MIXED_PRECISION else None) start_epoch = checkpoint_epoch + 1 else: start_epoch = 0 cu.load_checkpoint(cfg.TRAIN.CHECKPOINT_FILE_PATH, model) # Create the video train and val loaders. train_loader = loader.construct_loader(cfg, "train") val_loader = loader.construct_loader(cfg, "val") precise_bn_loader = ( loader.construct_loader(cfg, "train", is_precise_bn=True) if cfg.BN.USE_PRECISE_STATS else None ) # Create meters. if cfg.DETECTION.ENABLE: train_meter = AVAMeter(len(train_loader), cfg, mode="train") val_meter = AVAMeter(len(val_loader), cfg, mode="val") else: train_meter = TrainMeter(len(train_loader), cfg) val_meter = ValMeter(len(val_loader), cfg) # set up writer for logging to Tensorboard format. if cfg.TENSORBOARD.ENABLE and du.is_master_proc( cfg.NUM_GPUS * cfg.NUM_SHARDS ): writer = tb.TensorboardWriter(cfg) else: writer = None # Perform the training loop. logger.info("Start epoch: {}".format(start_epoch + 1)) total_epochs = cfg.SOLVER.MAX_EPOCH epoch_timer = EpochTimer() for cur_epoch in range(start_epoch, cfg.SOLVER.MAX_EPOCH): if cfg.MULTIGRID.LONG_CYCLE: cfg, changed = multigrid.update_long_cycle(cfg, cur_epoch) if changed: ( model, optimizer, train_loader, val_loader, precise_bn_loader, train_meter, val_meter, ) = build_trainer(cfg) # Load checkpoint. if cu.has_checkpoint(cfg.OUTPUT_DIR): last_checkpoint = cu.get_last_checkpoint(cfg.OUTPUT_DIR) assert "{:05d}.pyth".format(cur_epoch) in last_checkpoint else: last_checkpoint = cfg.TRAIN.CHECKPOINT_FILE_PATH logger.info("Load from {}".format(last_checkpoint)) cu.load_checkpoint( last_checkpoint, model, cfg.NUM_GPUS > 1, optimizer ) # Shuffle the dataset. loader.shuffle_dataset(train_loader, cur_epoch) # Train for one epoch. epoch_timer.epoch_tic() train_epoch( train_loader, model, loss_fun, optimizer, scaler, train_meter, cur_epoch, total_epochs, cfg, writer, ) epoch_timer.epoch_toc() logger.info( f"Epoch {cur_epoch} takes {epoch_timer.last_epoch_time():.2f}s. Epochs " f"from {start_epoch} to {cur_epoch} take " f"{epoch_timer.avg_epoch_time():.2f}s in average and " f"{epoch_timer.median_epoch_time():.2f}s in median." ) logger.info( f"For epoch {cur_epoch}, each iteraction takes " f"{epoch_timer.last_epoch_time()/len(train_loader):.2f}s in average. " f"From epoch {start_epoch} to {cur_epoch}, each iteraction takes " f"{epoch_timer.avg_epoch_time()/len(train_loader):.2f}s in average." ) is_checkp_epoch = cu.is_checkpoint_epoch( cfg, cur_epoch, None if multigrid is None else multigrid.schedule, ) is_eval_epoch = misc.is_eval_epoch( cfg, cur_epoch, None if multigrid is None else multigrid.schedule ) # Compute precise BN stats. if ( (is_checkp_epoch or is_eval_epoch) and cfg.BN.USE_PRECISE_STATS and len(get_bn_modules(model)) > 0 ): calculate_and_update_precise_bn( precise_bn_loader, model, min(cfg.BN.NUM_BATCHES_PRECISE, len(precise_bn_loader)), cfg.NUM_GPUS > 0, ) _ = misc.aggregate_sub_bn_stats(model) # Save a checkpoint. if is_checkp_epoch: cu.save_checkpoint( cfg.OUTPUT_DIR, model, optimizer, cur_epoch, cfg, scaler if cfg.TRAIN.MIXED_PRECISION else None, ) # Evaluate the model on validation set. if is_eval_epoch: eval_epoch(val_loader, model, val_meter, cur_epoch, cfg, writer) if writer is not None: writer.close()
21,443
34.562189
116
py
STTS
STTS-main/MViT/slowfast/__init__.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. from slowfast.utils.env import setup_environment setup_environment()
166
22.857143
71
py
STTS
STTS-main/MViT/slowfast/config/custom_config.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Add custom configs and default values""" def add_custom_config(_C): # Add your own customized configs. pass
217
20.8
71
py
STTS
STTS-main/MViT/slowfast/config/defaults.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Configs.""" from fvcore.common.config import CfgNode from . import custom_config # ----------------------------------------------------------------------------- # Config definition # ----------------------------------------------------------------------------- _C = CfgNode() # ---------------------------------------------------------------------------- # # Batch norm options # ---------------------------------------------------------------------------- # _C.BN = CfgNode() # Precise BN stats. _C.BN.USE_PRECISE_STATS = False # Number of samples use to compute precise bn. _C.BN.NUM_BATCHES_PRECISE = 200 # Weight decay value that applies on BN. _C.BN.WEIGHT_DECAY = 0.0 # Norm type, options include `batchnorm`, `sub_batchnorm`, `sync_batchnorm` _C.BN.NORM_TYPE = "batchnorm" # Parameter for SubBatchNorm, where it splits the batch dimension into # NUM_SPLITS splits, and run BN on each of them separately independently. _C.BN.NUM_SPLITS = 1 # Parameter for NaiveSyncBatchNorm3d, where the stats across `NUM_SYNC_DEVICES` # devices will be synchronized. _C.BN.NUM_SYNC_DEVICES = 1 # ---------------------------------------------------------------------------- # # Training options. # ---------------------------------------------------------------------------- # _C.TRAIN = CfgNode() # If True Train the model, else skip training. _C.TRAIN.ENABLE = True # Dataset. _C.TRAIN.DATASET = "kinetics" # Total mini-batch size. _C.TRAIN.BATCH_SIZE = 64 # Evaluate model on test data every eval period epochs. _C.TRAIN.EVAL_PERIOD = 10 # Save model checkpoint every checkpoint period epochs. _C.TRAIN.CHECKPOINT_PERIOD = 10 _C.TRAIN.TRAIN_TOPK_ONLY = False # Resume training from the latest checkpoint in the output directory. _C.TRAIN.AUTO_RESUME = True _C.TRAIN.FINETUNE = True # Path to the checkpoint to load the initial weight. _C.TRAIN.CHECKPOINT_FILE_PATH = "" # Checkpoint types include `caffe2` or `pytorch`. _C.TRAIN.CHECKPOINT_TYPE = "pytorch" # If True, perform inflation when loading checkpoint. _C.TRAIN.CHECKPOINT_INFLATE = False # If True, reset epochs when loading checkpoint. _C.TRAIN.CHECKPOINT_EPOCH_RESET = False # If set, clear all layer names according to the pattern provided. _C.TRAIN.CHECKPOINT_CLEAR_NAME_PATTERN = () # ("backbone.",) # If True, use FP16 for activations _C.TRAIN.MIXED_PRECISION = False # ---------------------------------------------------------------------------- # # Augmentation options. # ---------------------------------------------------------------------------- # _C.AUG = CfgNode() # Whether to enable randaug. _C.AUG.ENABLE = False # Number of repeated augmentations to used during training. # If this is greater than 1, then the actual batch size is # TRAIN.BATCH_SIZE * AUG.NUM_SAMPLE. _C.AUG.NUM_SAMPLE = 1 # Not used if using randaug. _C.AUG.COLOR_JITTER = 0.4 # RandAug parameters. _C.AUG.AA_TYPE = "rand-m9-mstd0.5-inc1" # Interpolation method. _C.AUG.INTERPOLATION = "bicubic" # Probability of random erasing. _C.AUG.RE_PROB = 0.25 # Random erasing mode. _C.AUG.RE_MODE = "pixel" # Random erase count. _C.AUG.RE_COUNT = 1 # Do not random erase first (clean) augmentation split. _C.AUG.RE_SPLIT = False # ---------------------------------------------------------------------------- # # MipUp options. # ---------------------------------------------------------------------------- # _C.MIXUP = CfgNode() # Whether to use mixup. _C.MIXUP.ENABLE = False # Mixup alpha. _C.MIXUP.ALPHA = 0.8 # Cutmix alpha. _C.MIXUP.CUTMIX_ALPHA = 1.0 # Probability of performing mixup or cutmix when either/both is enabled. _C.MIXUP.PROB = 1.0 # Probability of switching to cutmix when both mixup and cutmix enabled. _C.MIXUP.SWITCH_PROB = 0.5 # Label smoothing. _C.MIXUP.LABEL_SMOOTH_VALUE = 0.1 # ---------------------------------------------------------------------------- # # Testing options # ---------------------------------------------------------------------------- # _C.TEST = CfgNode() # If True test the model, else skip the testing. _C.TEST.ENABLE = True # Dataset for testing. _C.TEST.DATASET = "kinetics" # Total mini-batch size _C.TEST.BATCH_SIZE = 8 # Path to the checkpoint to load the initial weight. _C.TEST.CHECKPOINT_FILE_PATH = "" # Number of clips to sample from a video uniformly for aggregating the # prediction results. _C.TEST.NUM_ENSEMBLE_VIEWS = 10 # Number of crops to sample from a frame spatially for aggregating the # prediction results. _C.TEST.NUM_SPATIAL_CROPS = 3 _C.TEST.SUBSET = "full" # Checkpoint types include `caffe2` or `pytorch`. _C.TEST.CHECKPOINT_TYPE = "pytorch" # Path to saving prediction results file. _C.TEST.SAVE_RESULTS_PATH = "" # ----------------------------------------------------------------------------- # ResNet options # ----------------------------------------------------------------------------- _C.RESNET = CfgNode() # Transformation function. _C.RESNET.TRANS_FUNC = "bottleneck_transform" # Number of groups. 1 for ResNet, and larger than 1 for ResNeXt). _C.RESNET.NUM_GROUPS = 1 # Width of each group (64 -> ResNet; 4 -> ResNeXt). _C.RESNET.WIDTH_PER_GROUP = 64 # Apply relu in a inplace manner. _C.RESNET.INPLACE_RELU = True # Apply stride to 1x1 conv. _C.RESNET.STRIDE_1X1 = False # If true, initialize the gamma of the final BN of each block to zero. _C.RESNET.ZERO_INIT_FINAL_BN = False # Number of weight layers. _C.RESNET.DEPTH = 50 # If the current block has more than NUM_BLOCK_TEMP_KERNEL blocks, use temporal # kernel of 1 for the rest of the blocks. _C.RESNET.NUM_BLOCK_TEMP_KERNEL = [[3], [4], [6], [3]] # Size of stride on different res stages. _C.RESNET.SPATIAL_STRIDES = [[1], [2], [2], [2]] # Size of dilation on different res stages. _C.RESNET.SPATIAL_DILATIONS = [[1], [1], [1], [1]] # ---------------------------------------------------------------------------- # # X3D options # See https://arxiv.org/abs/2004.04730 for details about X3D Networks. # ---------------------------------------------------------------------------- # _C.X3D = CfgNode() # Width expansion factor. _C.X3D.WIDTH_FACTOR = 1.0 # Depth expansion factor. _C.X3D.DEPTH_FACTOR = 1.0 # Bottleneck expansion factor for the 3x3x3 conv. _C.X3D.BOTTLENECK_FACTOR = 1.0 # # Dimensions of the last linear layer before classificaiton. _C.X3D.DIM_C5 = 2048 # Dimensions of the first 3x3 conv layer. _C.X3D.DIM_C1 = 12 # Whether to scale the width of Res2, default is false. _C.X3D.SCALE_RES2 = False # Whether to use a BatchNorm (BN) layer before the classifier, default is false. _C.X3D.BN_LIN5 = False # Whether to use channelwise (=depthwise) convolution in the center (3x3x3) # convolution operation of the residual blocks. _C.X3D.CHANNELWISE_3x3x3 = True # ----------------------------------------------------------------------------- # Nonlocal options # ----------------------------------------------------------------------------- _C.NONLOCAL = CfgNode() # Index of each stage and block to add nonlocal layers. _C.NONLOCAL.LOCATION = [[[]], [[]], [[]], [[]]] # Number of group for nonlocal for each stage. _C.NONLOCAL.GROUP = [[1], [1], [1], [1]] # Instatiation to use for non-local layer. _C.NONLOCAL.INSTANTIATION = "dot_product" # Size of pooling layers used in Non-Local. _C.NONLOCAL.POOL = [ # Res2 [[1, 2, 2], [1, 2, 2]], # Res3 [[1, 2, 2], [1, 2, 2]], # Res4 [[1, 2, 2], [1, 2, 2]], # Res5 [[1, 2, 2], [1, 2, 2]], ] # ----------------------------------------------------------------------------- # Model options # ----------------------------------------------------------------------------- _C.MODEL = CfgNode() # Model architecture. _C.MODEL.ARCH = "slowfast" # Model name _C.MODEL.MODEL_NAME = "SlowFast" # The number of classes to predict for the model. _C.MODEL.NUM_CLASSES = 400 # Loss function. _C.MODEL.LOSS_FUNC = "cross_entropy" # Model architectures that has one single pathway. _C.MODEL.SINGLE_PATHWAY_ARCH = ["2d", "c2d", "i3d", "slow", "x3d", "mvit"] # Model architectures that has multiple pathways. _C.MODEL.MULTI_PATHWAY_ARCH = ["slowfast"] # Dropout rate before final projection in the backbone. _C.MODEL.DROPOUT_RATE = 0.5 # Randomly drop rate for Res-blocks, linearly increase from res2 to res5 _C.MODEL.DROPCONNECT_RATE = 0.0 # The std to initialize the fc layer(s). _C.MODEL.FC_INIT_STD = 0.01 # Activation layer for the output head. _C.MODEL.HEAD_ACT = "softmax" # Activation checkpointing enabled or not to save GPU memory. _C.MODEL.ACT_CHECKPOINT = False # ----------------------------------------------------------------------------- # MViT options # ----------------------------------------------------------------------------- _C.MVIT = CfgNode() # Options include `conv`, `max`. _C.MVIT.MODE = "conv" # If True, perform pool before projection in attention. _C.MVIT.POOL_FIRST = False # If True, use cls embed in the network, otherwise don't use cls_embed in transformer. _C.MVIT.CLS_EMBED_ON = True # Kernel size for patchtification. _C.MVIT.PATCH_KERNEL = [3, 7, 7] # Stride size for patchtification. _C.MVIT.PATCH_STRIDE = [2, 4, 4] _C.MVIT.DECAY_SIGMA = True _C.MVIT.SIGMA = 0.05 _C.MVIT.TIME_PRUNING_LOC = None _C.MVIT.SPACE_PRUNING_LOC = None _C.MVIT.TIME_SCORE = 'tpool' _C.MVIT.SPACE_SCORE = 'spatch' _C.MVIT.TIME_LEFT_RATIO = [0.5] _C.MVIT.SPACE_LEFT_RATIO = [0.5105] # Padding size for patchtification. _C.MVIT.PATCH_PADDING = [2, 4, 4] # If True, use 2d patch, otherwise use 3d patch. _C.MVIT.PATCH_2D = False # Base embedding dimension for the transformer. _C.MVIT.EMBED_DIM = 96 # Base num of heads for the transformer. _C.MVIT.NUM_HEADS = 1 # Dimension reduction ratio for the MLP layers. _C.MVIT.MLP_RATIO = 4.0 # If use, use bias term in attention fc layers. _C.MVIT.QKV_BIAS = True # Drop path rate for the tranfomer. _C.MVIT.DROPPATH_RATE = 0.1 # Depth of the transformer. _C.MVIT.DEPTH = 16 # Normalization layer for the transformer. Only layernorm is supported now. _C.MVIT.NORM = "layernorm" # Dimension multiplication at layer i. If 2.0 is used, then the next block will increase # the dimension by 2 times. Format: [depth_i: mul_dim_ratio] _C.MVIT.DIM_MUL = [] # Head number multiplication at layer i. If 2.0 is used, then the next block will # increase the number of heads by 2 times. Format: [depth_i: head_mul_ratio] _C.MVIT.HEAD_MUL = [] # Stride size for the Pool KV at layer i. # Format: [[i, stride_t_i, stride_h_i, stride_w_i], ...,] _C.MVIT.POOL_KV_STRIDE = None # Initial stride size for KV at layer 1. The stride size will be further reduced with # the raio of MVIT.DIM_MUL. If will overwrite MVIT.POOL_KV_STRIDE if not None. _C.MVIT.POOL_KV_STRIDE_ADAPTIVE = None # Stride size for the Pool Q at layer i. # Format: [[i, stride_t_i, stride_h_i, stride_w_i], ...,] _C.MVIT.POOL_Q_STRIDE = [] # If not None, overwrite the KV_KERNEL and Q_KERNEL size with POOL_KVQ_CONV_SIZ. # Otherwise the kernel_size is [s + 1 if s > 1 else s for s in stride_size]. _C.MVIT.POOL_KVQ_KERNEL = None # If True, perform no decay on positional embedding and cls embedding. _C.MVIT.ZERO_DECAY_POS_CLS = True # If True, use norm after stem. _C.MVIT.NORM_STEM = False # If True, perform separate positional embedding. _C.MVIT.SEP_POS_EMBED = False # Dropout rate for the MViT backbone. _C.MVIT.DROPOUT_RATE = 0.0 # ----------------------------------------------------------------------------- # SlowFast options # ----------------------------------------------------------------------------- _C.SLOWFAST = CfgNode() # Corresponds to the inverse of the channel reduction ratio, $\beta$ between # the Slow and Fast pathways. _C.SLOWFAST.BETA_INV = 8 # Corresponds to the frame rate reduction ratio, $\alpha$ between the Slow and # Fast pathways. _C.SLOWFAST.ALPHA = 8 # Ratio of channel dimensions between the Slow and Fast pathways. _C.SLOWFAST.FUSION_CONV_CHANNEL_RATIO = 2 # Kernel dimension used for fusing information from Fast pathway to Slow # pathway. _C.SLOWFAST.FUSION_KERNEL_SZ = 5 # ----------------------------------------------------------------------------- # Data options # ----------------------------------------------------------------------------- _C.DATA = CfgNode() # The path to the data directory. _C.DATA.PATH_TO_DATA_DIR = "" # The separator used between path and label. _C.DATA.PATH_LABEL_SEPARATOR = "," # Video path prefix if any. _C.DATA.PATH_PREFIX = "" # The number of frames of the input clip. _C.DATA.NUM_FRAMES = 8 # The video sampling rate of the input clip. _C.DATA.SAMPLING_RATE = 8 # Eigenvalues for PCA jittering. Note PCA is RGB based. _C.DATA.TRAIN_PCA_EIGVAL = [0.225, 0.224, 0.229] # Eigenvectors for PCA jittering. _C.DATA.TRAIN_PCA_EIGVEC = [ [-0.5675, 0.7192, 0.4009], [-0.5808, -0.0045, -0.8140], [-0.5836, -0.6948, 0.4203], ] # If a imdb have been dumpped to a local file with the following format: # `{"im_path": im_path, "class": cont_id}` # then we can skip the construction of imdb and load it from the local file. _C.DATA.PATH_TO_PRELOAD_IMDB = "" # The mean value of the video raw pixels across the R G B channels. _C.DATA.MEAN = [0.45, 0.45, 0.45] # List of input frame channel dimensions. _C.DATA.INPUT_CHANNEL_NUM = [3, 3] # The std value of the video raw pixels across the R G B channels. _C.DATA.STD = [0.225, 0.225, 0.225] # The spatial augmentation jitter scales for training. _C.DATA.TRAIN_JITTER_SCALES = [256, 320] # The relative scale range of Inception-style area based random resizing augmentation. # If this is provided, DATA.TRAIN_JITTER_SCALES above is ignored. _C.DATA.TRAIN_JITTER_SCALES_RELATIVE = [] # The relative aspect ratio range of Inception-style area based random resizing # augmentation. _C.DATA.TRAIN_JITTER_ASPECT_RELATIVE = [] # If True, perform stride length uniform temporal sampling. _C.DATA.USE_OFFSET_SAMPLING = False # Whether to apply motion shift for augmentation. _C.DATA.TRAIN_JITTER_MOTION_SHIFT = False # The spatial crop size for training. _C.DATA.TRAIN_CROP_SIZE = 224 # The spatial crop size for testing. _C.DATA.TEST_CROP_SIZE = 256 # Input videos may has different fps, convert it to the target video fps before # frame sampling. _C.DATA.TARGET_FPS = 30 # Decoding backend, options include `pyav` or `torchvision` _C.DATA.DECODING_BACKEND = "pyav" # if True, sample uniformly in [1 / max_scale, 1 / min_scale] and take a # reciprocal to get the scale. If False, take a uniform sample from # [min_scale, max_scale]. _C.DATA.INV_UNIFORM_SAMPLE = False # If True, perform random horizontal flip on the video frames during training. _C.DATA.RANDOM_FLIP = True # If True, calculdate the map as metric. _C.DATA.MULTI_LABEL = False # Method to perform the ensemble, options include "sum" and "max". _C.DATA.ENSEMBLE_METHOD = "sum" # If True, revert the default input channel (RBG <-> BGR). _C.DATA.REVERSE_INPUT_CHANNEL = False # ---------------------------------------------------------------------------- # # Optimizer options # ---------------------------------------------------------------------------- # _C.SOLVER = CfgNode() # Base learning rate. _C.SOLVER.BASE_LR = 0.1 # Learning rate policy (see utils/lr_policy.py for options and examples). _C.SOLVER.LR_POLICY = "cosine" # Final learning rates for 'cosine' policy. _C.SOLVER.COSINE_END_LR = 0.0 # Exponential decay factor. _C.SOLVER.GAMMA = 0.1 _C.SOLVER.BACKBONE_LR = 0.01 # Step size for 'exp' and 'cos' policies (in epochs). _C.SOLVER.STEP_SIZE = 1 # Steps for 'steps_' policies (in epochs). _C.SOLVER.STEPS = [] # Learning rates for 'steps_' policies. _C.SOLVER.LRS = [] # Maximal number of epochs. _C.SOLVER.MAX_EPOCH = 300 # Momentum. _C.SOLVER.MOMENTUM = 0.9 # Momentum dampening. _C.SOLVER.DAMPENING = 0.0 # Nesterov momentum. _C.SOLVER.NESTEROV = True # L2 regularization. _C.SOLVER.WEIGHT_DECAY = 1e-4 # Start the warm up from SOLVER.BASE_LR * SOLVER.WARMUP_FACTOR. _C.SOLVER.WARMUP_FACTOR = 0.1 # Gradually warm up the SOLVER.BASE_LR over this number of epochs. _C.SOLVER.WARMUP_EPOCHS = 0.0 # The start learning rate of the warm up. _C.SOLVER.WARMUP_START_LR = 0.01 # Optimization method. _C.SOLVER.OPTIMIZING_METHOD = "sgd" # Base learning rate is linearly scaled with NUM_SHARDS. _C.SOLVER.BASE_LR_SCALE_NUM_SHARDS = False # If True, start from the peak cosine learning rate after warm up. _C.SOLVER.COSINE_AFTER_WARMUP = False # If True, perform no weight decay on parameter with one dimension (bias term, etc). _C.SOLVER.ZERO_WD_1D_PARAM = False # Clip gradient at this value before optimizer update _C.SOLVER.CLIP_GRAD_VAL = None # Clip gradient at this norm before optimizer update _C.SOLVER.CLIP_GRAD_L2NORM = None # ---------------------------------------------------------------------------- # # Misc options # ---------------------------------------------------------------------------- # # Number of GPUs to use (applies to both training and testing). _C.NUM_GPUS = 1 # Number of machine to use for the job. _C.NUM_SHARDS = 1 # The index of the current machine. _C.SHARD_ID = 0 # Output basedir. _C.OUTPUT_DIR = "./tmp" # Note that non-determinism may still be present due to non-deterministic # operator implementations in GPU operator libraries. _C.RNG_SEED = 1 # Log period in iters. _C.LOG_PERIOD = 10 # If True, log the model info. _C.LOG_MODEL_INFO = False # Distributed backend. _C.DIST_BACKEND = "nccl" # ---------------------------------------------------------------------------- # # Benchmark options # ---------------------------------------------------------------------------- # _C.BENCHMARK = CfgNode() # Number of epochs for data loading benchmark. _C.BENCHMARK.NUM_EPOCHS = 5 # Log period in iters for data loading benchmark. _C.BENCHMARK.LOG_PERIOD = 100 # If True, shuffle dataloader for epoch during benchmark. _C.BENCHMARK.SHUFFLE = True # ---------------------------------------------------------------------------- # # Common train/test data loader options # ---------------------------------------------------------------------------- # _C.DATA_LOADER = CfgNode() # Number of data loader workers per training process. _C.DATA_LOADER.NUM_WORKERS = 8 # Load data to pinned host memory. _C.DATA_LOADER.PIN_MEMORY = True # Enable multi thread decoding. _C.DATA_LOADER.ENABLE_MULTI_THREAD_DECODE = False # ---------------------------------------------------------------------------- # # Detection options. # ---------------------------------------------------------------------------- # _C.DETECTION = CfgNode() # Whether enable video detection. _C.DETECTION.ENABLE = False # Aligned version of RoI. More details can be found at slowfast/models/head_helper.py _C.DETECTION.ALIGNED = True # Spatial scale factor. _C.DETECTION.SPATIAL_SCALE_FACTOR = 16 # RoI tranformation resolution. _C.DETECTION.ROI_XFORM_RESOLUTION = 7 # ----------------------------------------------------------------------------- # AVA Dataset options # ----------------------------------------------------------------------------- _C.AVA = CfgNode() # Directory path of frames. _C.AVA.FRAME_DIR = "/mnt/fair-flash3-east/ava_trainval_frames.img/" # Directory path for files of frame lists. _C.AVA.FRAME_LIST_DIR = ( "/mnt/vol/gfsai-flash3-east/ai-group/users/haoqifan/ava/frame_list/" ) # Directory path for annotation files. _C.AVA.ANNOTATION_DIR = ( "/mnt/vol/gfsai-flash3-east/ai-group/users/haoqifan/ava/frame_list/" ) # Filenames of training samples list files. _C.AVA.TRAIN_LISTS = ["train.csv"] # Filenames of test samples list files. _C.AVA.TEST_LISTS = ["val.csv"] # Filenames of box list files for training. Note that we assume files which # contains predicted boxes will have a suffix "predicted_boxes" in the # filename. _C.AVA.TRAIN_GT_BOX_LISTS = ["ava_train_v2.2.csv"] _C.AVA.TRAIN_PREDICT_BOX_LISTS = [] # Filenames of box list files for test. _C.AVA.TEST_PREDICT_BOX_LISTS = ["ava_val_predicted_boxes.csv"] # This option controls the score threshold for the predicted boxes to use. _C.AVA.DETECTION_SCORE_THRESH = 0.9 # If use BGR as the format of input frames. _C.AVA.BGR = False # Training augmentation parameters # Whether to use color augmentation method. _C.AVA.TRAIN_USE_COLOR_AUGMENTATION = False # Whether to only use PCA jitter augmentation when using color augmentation # method (otherwise combine with color jitter method). _C.AVA.TRAIN_PCA_JITTER_ONLY = True # Whether to do horizontal flipping during test. _C.AVA.TEST_FORCE_FLIP = False # Whether to use full test set for validation split. _C.AVA.FULL_TEST_ON_VAL = False # The name of the file to the ava label map. _C.AVA.LABEL_MAP_FILE = "ava_action_list_v2.2_for_activitynet_2019.pbtxt" # The name of the file to the ava exclusion. _C.AVA.EXCLUSION_FILE = "ava_val_excluded_timestamps_v2.2.csv" # The name of the file to the ava groundtruth. _C.AVA.GROUNDTRUTH_FILE = "ava_val_v2.2.csv" # Backend to process image, includes `pytorch` and `cv2`. _C.AVA.IMG_PROC_BACKEND = "cv2" # ---------------------------------------------------------------------------- # # Multigrid training options # See https://arxiv.org/abs/1912.00998 for details about multigrid training. # ---------------------------------------------------------------------------- # _C.MULTIGRID = CfgNode() # Multigrid training allows us to train for more epochs with fewer iterations. # This hyperparameter specifies how many times more epochs to train. # The default setting in paper trains for 1.5x more epochs than baseline. _C.MULTIGRID.EPOCH_FACTOR = 1.5 # Enable short cycles. _C.MULTIGRID.SHORT_CYCLE = False # Short cycle additional spatial dimensions relative to the default crop size. _C.MULTIGRID.SHORT_CYCLE_FACTORS = [0.5, 0.5 ** 0.5] _C.MULTIGRID.LONG_CYCLE = False # (Temporal, Spatial) dimensions relative to the default shape. _C.MULTIGRID.LONG_CYCLE_FACTORS = [ (0.25, 0.5 ** 0.5), (0.5, 0.5 ** 0.5), (0.5, 1), (1, 1), ] # While a standard BN computes stats across all examples in a GPU, # for multigrid training we fix the number of clips to compute BN stats on. # See https://arxiv.org/abs/1912.00998 for details. _C.MULTIGRID.BN_BASE_SIZE = 8 # Multigrid training epochs are not proportional to actual training time or # computations, so _C.TRAIN.EVAL_PERIOD leads to too frequent or rare # evaluation. We use a multigrid-specific rule to determine when to evaluate: # This hyperparameter defines how many times to evaluate a model per long # cycle shape. _C.MULTIGRID.EVAL_FREQ = 3 # No need to specify; Set automatically and used as global variables. _C.MULTIGRID.LONG_CYCLE_SAMPLING_RATE = 0 _C.MULTIGRID.DEFAULT_B = 0 _C.MULTIGRID.DEFAULT_T = 0 _C.MULTIGRID.DEFAULT_S = 0 # ----------------------------------------------------------------------------- # Tensorboard Visualization Options # ----------------------------------------------------------------------------- _C.TENSORBOARD = CfgNode() # Log to summary writer, this will automatically. # log loss, lr and metrics during train/eval. _C.TENSORBOARD.ENABLE = False # Provide path to prediction results for visualization. # This is a pickle file of [prediction_tensor, label_tensor] _C.TENSORBOARD.PREDICTIONS_PATH = "" # Path to directory for tensorboard logs. # Default to to cfg.OUTPUT_DIR/runs-{cfg.TRAIN.DATASET}. _C.TENSORBOARD.LOG_DIR = "" # Path to a json file providing class_name - id mapping # in the format {"class_name1": id1, "class_name2": id2, ...}. # This file must be provided to enable plotting confusion matrix # by a subset or parent categories. _C.TENSORBOARD.CLASS_NAMES_PATH = "" # Path to a json file for categories -> classes mapping # in the format {"parent_class": ["child_class1", "child_class2",...], ...}. _C.TENSORBOARD.CATEGORIES_PATH = "" # Config for confusion matrices visualization. _C.TENSORBOARD.CONFUSION_MATRIX = CfgNode() # Visualize confusion matrix. _C.TENSORBOARD.CONFUSION_MATRIX.ENABLE = False # Figure size of the confusion matrices plotted. _C.TENSORBOARD.CONFUSION_MATRIX.FIGSIZE = [8, 8] # Path to a subset of categories to visualize. # File contains class names separated by newline characters. _C.TENSORBOARD.CONFUSION_MATRIX.SUBSET_PATH = "" # Config for histogram visualization. _C.TENSORBOARD.HISTOGRAM = CfgNode() # Visualize histograms. _C.TENSORBOARD.HISTOGRAM.ENABLE = False # Path to a subset of classes to plot histograms. # Class names must be separated by newline characters. _C.TENSORBOARD.HISTOGRAM.SUBSET_PATH = "" # Visualize top-k most predicted classes on histograms for each # chosen true label. _C.TENSORBOARD.HISTOGRAM.TOPK = 10 # Figure size of the histograms plotted. _C.TENSORBOARD.HISTOGRAM.FIGSIZE = [8, 8] # Config for layers' weights and activations visualization. # _C.TENSORBOARD.ENABLE must be True. _C.TENSORBOARD.MODEL_VIS = CfgNode() # If False, skip model visualization. _C.TENSORBOARD.MODEL_VIS.ENABLE = False # If False, skip visualizing model weights. _C.TENSORBOARD.MODEL_VIS.MODEL_WEIGHTS = False # If False, skip visualizing model activations. _C.TENSORBOARD.MODEL_VIS.ACTIVATIONS = False # If False, skip visualizing input videos. _C.TENSORBOARD.MODEL_VIS.INPUT_VIDEO = False # List of strings containing data about layer names and their indexing to # visualize weights and activations for. The indexing is meant for # choosing a subset of activations outputed by a layer for visualization. # If indexing is not specified, visualize all activations outputed by the layer. # For each string, layer name and indexing is separated by whitespaces. # e.g.: [layer1 1,2;1,2, layer2, layer3 150,151;3,4]; this means for each array `arr` # along the batch dimension in `layer1`, we take arr[[1, 2], [1, 2]] _C.TENSORBOARD.MODEL_VIS.LAYER_LIST = [] # Top-k predictions to plot on videos _C.TENSORBOARD.MODEL_VIS.TOPK_PREDS = 1 # Colormap to for text boxes and bounding boxes colors _C.TENSORBOARD.MODEL_VIS.COLORMAP = "Pastel2" # Config for visualization video inputs with Grad-CAM. # _C.TENSORBOARD.ENABLE must be True. _C.TENSORBOARD.MODEL_VIS.GRAD_CAM = CfgNode() # Whether to run visualization using Grad-CAM technique. _C.TENSORBOARD.MODEL_VIS.GRAD_CAM.ENABLE = True # CNN layers to use for Grad-CAM. The number of layers must be equal to # number of pathway(s). _C.TENSORBOARD.MODEL_VIS.GRAD_CAM.LAYER_LIST = [] # If True, visualize Grad-CAM using true labels for each instances. # If False, use the highest predicted class. _C.TENSORBOARD.MODEL_VIS.GRAD_CAM.USE_TRUE_LABEL = False # Colormap to for text boxes and bounding boxes colors _C.TENSORBOARD.MODEL_VIS.GRAD_CAM.COLORMAP = "viridis" # Config for visualization for wrong prediction visualization. # _C.TENSORBOARD.ENABLE must be True. _C.TENSORBOARD.WRONG_PRED_VIS = CfgNode() _C.TENSORBOARD.WRONG_PRED_VIS.ENABLE = False # Folder tag to origanize model eval videos under. _C.TENSORBOARD.WRONG_PRED_VIS.TAG = "Incorrectly classified videos." # Subset of labels to visualize. Only wrong predictions with true labels # within this subset is visualized. _C.TENSORBOARD.WRONG_PRED_VIS.SUBSET_PATH = "" _C.USE_MINI = False # ---------------------------------------------------------------------------- # # Demo options # ---------------------------------------------------------------------------- # _C.DEMO = CfgNode() # Run model in DEMO mode. _C.DEMO.ENABLE = False # Path to a json file providing class_name - id mapping # in the format {"class_name1": id1, "class_name2": id2, ...}. _C.DEMO.LABEL_FILE_PATH = "" # Specify a camera device as input. This will be prioritized # over input video if set. # If -1, use input video instead. _C.DEMO.WEBCAM = -1 # Path to input video for demo. _C.DEMO.INPUT_VIDEO = "" # Custom width for reading input video data. _C.DEMO.DISPLAY_WIDTH = 0 # Custom height for reading input video data. _C.DEMO.DISPLAY_HEIGHT = 0 # Path to Detectron2 object detection model configuration, # only used for detection tasks. _C.DEMO.DETECTRON2_CFG = "COCO-Detection/faster_rcnn_R_50_FPN_3x.yaml" # Path to Detectron2 object detection model pre-trained weights. _C.DEMO.DETECTRON2_WEIGHTS = "detectron2://COCO-Detection/faster_rcnn_R_50_FPN_3x/137849458/model_final_280758.pkl" # Threshold for choosing predicted bounding boxes by Detectron2. _C.DEMO.DETECTRON2_THRESH = 0.9 # Number of overlapping frames between 2 consecutive clips. # Increase this number for more frequent action predictions. # The number of overlapping frames cannot be larger than # half of the sequence length `cfg.DATA.NUM_FRAMES * cfg.DATA.SAMPLING_RATE` _C.DEMO.BUFFER_SIZE = 0 # If specified, the visualized outputs will be written this a video file of # this path. Otherwise, the visualized outputs will be displayed in a window. _C.DEMO.OUTPUT_FILE = "" # Frames per second rate for writing to output video file. # If not set (-1), use fps rate from input file. _C.DEMO.OUTPUT_FPS = -1 # Input format from demo video reader ("RGB" or "BGR"). _C.DEMO.INPUT_FORMAT = "BGR" # Draw visualization frames in [keyframe_idx - CLIP_VIS_SIZE, keyframe_idx + CLIP_VIS_SIZE] inclusively. _C.DEMO.CLIP_VIS_SIZE = 10 # Number of processes to run video visualizer. _C.DEMO.NUM_VIS_INSTANCES = 2 # Path to pre-computed predicted boxes _C.DEMO.PREDS_BOXES = "" # Whether to run in with multi-threaded video reader. _C.DEMO.THREAD_ENABLE = False # Take one clip for every `DEMO.NUM_CLIPS_SKIP` + 1 for prediction and visualization. # This is used for fast demo speed by reducing the prediction/visualiztion frequency. # If -1, take the most recent read clip for visualization. This mode is only supported # if `DEMO.THREAD_ENABLE` is set to True. _C.DEMO.NUM_CLIPS_SKIP = 0 # Path to ground-truth boxes and labels (optional) _C.DEMO.GT_BOXES = "" # The starting second of the video w.r.t bounding boxes file. _C.DEMO.STARTING_SECOND = 900 # Frames per second of the input video/folder of images. _C.DEMO.FPS = 30 # Visualize with top-k predictions or predictions above certain threshold(s). # Option: {"thres", "top-k"} _C.DEMO.VIS_MODE = "thres" # Threshold for common class names. _C.DEMO.COMMON_CLASS_THRES = 0.7 # Theshold for uncommon class names. This will not be # used if `_C.DEMO.COMMON_CLASS_NAMES` is empty. _C.DEMO.UNCOMMON_CLASS_THRES = 0.3 # This is chosen based on distribution of examples in # each classes in AVA dataset. _C.DEMO.COMMON_CLASS_NAMES = [ "watch (a person)", "talk to (e.g., self, a person, a group)", "listen to (a person)", "touch (an object)", "carry/hold (an object)", "walk", "sit", "lie/sleep", "bend/bow (at the waist)", ] # Slow-motion rate for the visualization. The visualized portions of the # video will be played `_C.DEMO.SLOWMO` times slower than usual speed. _C.DEMO.SLOWMO = 1 # Add custom config with default values. custom_config.add_custom_config(_C) def assert_and_infer_cfg(cfg): # BN assertions. if cfg.BN.USE_PRECISE_STATS: assert cfg.BN.NUM_BATCHES_PRECISE >= 0 # TRAIN assertions. assert cfg.TRAIN.CHECKPOINT_TYPE in ["pytorch", "caffe2"] assert cfg.NUM_GPUS == 0 or cfg.TRAIN.BATCH_SIZE % cfg.NUM_GPUS == 0 # TEST assertions. assert cfg.TEST.CHECKPOINT_TYPE in ["pytorch", "caffe2"] assert cfg.NUM_GPUS == 0 or cfg.TEST.BATCH_SIZE % cfg.NUM_GPUS == 0 # RESNET assertions. assert cfg.RESNET.NUM_GROUPS > 0 assert cfg.RESNET.WIDTH_PER_GROUP > 0 assert cfg.RESNET.WIDTH_PER_GROUP % cfg.RESNET.NUM_GROUPS == 0 # Execute LR scaling by num_shards. if cfg.SOLVER.BASE_LR_SCALE_NUM_SHARDS: cfg.SOLVER.BASE_LR *= cfg.NUM_SHARDS cfg.SOLVER.WARMUP_START_LR *= cfg.NUM_SHARDS cfg.SOLVER.COSINE_END_LR *= cfg.NUM_SHARDS # General assertions. assert cfg.SHARD_ID < cfg.NUM_SHARDS return cfg def get_cfg(): """ Get a copy of the default config. """ return _C.clone()
31,843
31.230769
115
py
STTS
STTS-main/MViT/slowfast/config/__init__.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved.
95
31
71
py
STTS
STTS-main/MViT/slowfast/models/operators.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Custom operators.""" import torch import torch.nn as nn class Swish(nn.Module): """Swish activation function: x * sigmoid(x).""" def __init__(self): super(Swish, self).__init__() def forward(self, x): return SwishEfficient.apply(x) class SwishEfficient(torch.autograd.Function): """Swish activation function: x * sigmoid(x).""" @staticmethod def forward(ctx, x): result = x * torch.sigmoid(x) ctx.save_for_backward(x) return result @staticmethod def backward(ctx, grad_output): x = ctx.saved_variables[0] sigmoid_x = torch.sigmoid(x) return grad_output * (sigmoid_x * (1 + x * (1 - sigmoid_x))) class SE(nn.Module): """Squeeze-and-Excitation (SE) block w/ Swish: AvgPool, FC, Swish, FC, Sigmoid.""" def _round_width(self, width, multiplier, min_width=8, divisor=8): """ Round width of filters based on width multiplier Args: width (int): the channel dimensions of the input. multiplier (float): the multiplication factor. min_width (int): the minimum width after multiplication. divisor (int): the new width should be dividable by divisor. """ if not multiplier: return width width *= multiplier min_width = min_width or divisor width_out = max( min_width, int(width + divisor / 2) // divisor * divisor ) if width_out < 0.9 * width: width_out += divisor return int(width_out) def __init__(self, dim_in, ratio, relu_act=True): """ Args: dim_in (int): the channel dimensions of the input. ratio (float): the channel reduction ratio for squeeze. relu_act (bool): whether to use ReLU activation instead of Swish (default). divisor (int): the new width should be dividable by divisor. """ super(SE, self).__init__() self.avg_pool = nn.AdaptiveAvgPool3d((1, 1, 1)) dim_fc = self._round_width(dim_in, ratio) self.fc1 = nn.Conv3d(dim_in, dim_fc, 1, bias=True) self.fc1_act = nn.ReLU() if relu_act else Swish() self.fc2 = nn.Conv3d(dim_fc, dim_in, 1, bias=True) self.fc2_sig = nn.Sigmoid() def forward(self, x): x_in = x for module in self.children(): x = module(x) return x_in * x
2,552
29.759036
86
py
STTS
STTS-main/MViT/slowfast/models/custom_video_model_builder.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """A More Flexible Video models."""
133
21.333333
71
py
STTS
STTS-main/MViT/slowfast/models/losses.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Loss functions.""" import torch import torch.nn as nn import torch.nn.functional as F from slowfast.models.topk import batched_index_select class SoftTargetCrossEntropy(nn.Module): """ Cross entropy loss with soft target. """ def __init__(self, reduction="mean"): """ Args: reduction (str): specifies reduction to apply to the output. It can be "mean" (default) or "none". """ super(SoftTargetCrossEntropy, self).__init__() self.reduction = reduction def forward(self, x, y): loss = torch.sum(-y * F.log_softmax(x, dim=-1), dim=-1) if self.reduction == "mean": return loss.mean() elif self.reduction == "none": return loss else: raise NotImplementedError class SoftTargetCrossEntropyPruning(nn.Module): """ Cross entropy loss with soft target. """ def __init__(self, ratio_weight=2.0, pruning_loc=[0], keep_ratio=[0.5], clf_weight=1.0, reduction="mean"): """ Args: reduction (str): specifies reduction to apply to the output. It can be "mean" (default) or "none". """ super(SoftTargetCrossEntropyPruning, self).__init__() self.reduction = reduction self.clf_weight = clf_weight self.pruning_loc = pruning_loc self.keep_ratio = keep_ratio self.cls_loss = 0 self.ratio_loss = 0 self.ratio_weight = ratio_weight def forward(self, x, y): pred, out_pred_score = x cls_loss = torch.sum(-y * F.log_softmax(pred, dim=-1), dim=-1) if self.reduction == "mean": cls_loss = cls_loss.mean() elif self.reduction == "none": cls_loss = cls_loss else: raise NotImplementedError pred_loss = 0.0 ratio = self.keep_ratio left_ratio = 1. for i, score in enumerate(out_pred_score): pos_ratio = score.mean(1) left_ratio = left_ratio * ratio[i] print(left_ratio, pos_ratio) pred_loss = pred_loss + ((pos_ratio - left_ratio) ** 2).mean() loss = self.clf_weight * cls_loss + self.ratio_weight * pred_loss / len(self.pruning_loc) return loss _LOSSES = { "cross_entropy": nn.CrossEntropyLoss, "bce": nn.BCELoss, "bce_logit": nn.BCEWithLogitsLoss, "soft_cross_entropy": SoftTargetCrossEntropy, "soft_cross_entropy_pruning": SoftTargetCrossEntropyPruning, } def get_loss_func(loss_name): """ Retrieve the loss given the loss name. Args (int): loss_name: the name of the loss to use. """ if loss_name not in _LOSSES.keys(): raise NotImplementedError("Loss {} is not supported".format(loss_name)) return _LOSSES[loss_name]
2,960
28.61
110
py
STTS
STTS-main/MViT/slowfast/models/batchnorm_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """BatchNorm (BN) utility functions and custom batch-size BN implementations""" from functools import partial import torch import torch.distributed as dist import torch.nn as nn from torch.autograd.function import Function import slowfast.utils.distributed as du def get_norm(cfg): """ Args: cfg (CfgNode): model building configs, details are in the comments of the config file. Returns: nn.Module: the normalization layer. """ if cfg.BN.NORM_TYPE == "batchnorm": return nn.BatchNorm3d elif cfg.BN.NORM_TYPE == "sub_batchnorm": return partial(SubBatchNorm3d, num_splits=cfg.BN.NUM_SPLITS) elif cfg.BN.NORM_TYPE == "sync_batchnorm": return partial( NaiveSyncBatchNorm3d, num_sync_devices=cfg.BN.NUM_SYNC_DEVICES ) else: raise NotImplementedError( "Norm type {} is not supported".format(cfg.BN.NORM_TYPE) ) class SubBatchNorm3d(nn.Module): """ The standard BN layer computes stats across all examples in a GPU. In some cases it is desirable to compute stats across only a subset of examples (e.g., in multigrid training https://arxiv.org/abs/1912.00998). SubBatchNorm3d splits the batch dimension into N splits, and run BN on each of them separately (so that the stats are computed on each subset of examples (1/N of batch) independently. During evaluation, it aggregates the stats from all splits into one BN. """ def __init__(self, num_splits, **args): """ Args: num_splits (int): number of splits. args (list): other arguments. """ super(SubBatchNorm3d, self).__init__() self.num_splits = num_splits num_features = args["num_features"] # Keep only one set of weight and bias. if args.get("affine", True): self.affine = True args["affine"] = False self.weight = torch.nn.Parameter(torch.ones(num_features)) self.bias = torch.nn.Parameter(torch.zeros(num_features)) else: self.affine = False self.bn = nn.BatchNorm3d(**args) args["num_features"] = num_features * num_splits self.split_bn = nn.BatchNorm3d(**args) def _get_aggregated_mean_std(self, means, stds, n): """ Calculate the aggregated mean and stds. Args: means (tensor): mean values. stds (tensor): standard deviations. n (int): number of sets of means and stds. """ mean = means.view(n, -1).sum(0) / n std = ( stds.view(n, -1).sum(0) / n + ((means.view(n, -1) - mean) ** 2).view(n, -1).sum(0) / n ) return mean.detach(), std.detach() def aggregate_stats(self): """ Synchronize running_mean, and running_var. Call this before eval. """ if self.split_bn.track_running_stats: ( self.bn.running_mean.data, self.bn.running_var.data, ) = self._get_aggregated_mean_std( self.split_bn.running_mean, self.split_bn.running_var, self.num_splits, ) def forward(self, x): if self.training: n, c, t, h, w = x.shape x = x.view(n // self.num_splits, c * self.num_splits, t, h, w) x = self.split_bn(x) x = x.view(n, c, t, h, w) else: x = self.bn(x) if self.affine: x = x * self.weight.view((-1, 1, 1, 1)) x = x + self.bias.view((-1, 1, 1, 1)) return x class GroupGather(Function): """ GroupGather performs all gather on each of the local process/ GPU groups. """ @staticmethod def forward(ctx, input, num_sync_devices, num_groups): """ Perform forwarding, gathering the stats across different process/ GPU group. """ ctx.num_sync_devices = num_sync_devices ctx.num_groups = num_groups input_list = [ torch.zeros_like(input) for k in range(du.get_local_size()) ] dist.all_gather( input_list, input, async_op=False, group=du._LOCAL_PROCESS_GROUP ) inputs = torch.stack(input_list, dim=0) if num_groups > 1: rank = du.get_local_rank() group_idx = rank // num_sync_devices inputs = inputs[ group_idx * num_sync_devices : (group_idx + 1) * num_sync_devices ] inputs = torch.sum(inputs, dim=0) return inputs @staticmethod def backward(ctx, grad_output): """ Perform backwarding, gathering the gradients across different process/ GPU group. """ grad_output_list = [ torch.zeros_like(grad_output) for k in range(du.get_local_size()) ] dist.all_gather( grad_output_list, grad_output, async_op=False, group=du._LOCAL_PROCESS_GROUP, ) grads = torch.stack(grad_output_list, dim=0) if ctx.num_groups > 1: rank = du.get_local_rank() group_idx = rank // ctx.num_sync_devices grads = grads[ group_idx * ctx.num_sync_devices : (group_idx + 1) * ctx.num_sync_devices ] grads = torch.sum(grads, dim=0) return grads, None, None class NaiveSyncBatchNorm3d(nn.BatchNorm3d): def __init__(self, num_sync_devices, **args): """ Naive version of Synchronized 3D BatchNorm. Args: num_sync_devices (int): number of device to sync. args (list): other arguments. """ self.num_sync_devices = num_sync_devices if self.num_sync_devices > 0: assert du.get_local_size() % self.num_sync_devices == 0, ( du.get_local_size(), self.num_sync_devices, ) self.num_groups = du.get_local_size() // self.num_sync_devices else: self.num_sync_devices = du.get_local_size() self.num_groups = 1 super(NaiveSyncBatchNorm3d, self).__init__(**args) def forward(self, input): if du.get_local_size() == 1 or not self.training: return super().forward(input) assert input.shape[0] > 0, "SyncBatchNorm does not support empty inputs" C = input.shape[1] mean = torch.mean(input, dim=[0, 2, 3, 4]) meansqr = torch.mean(input * input, dim=[0, 2, 3, 4]) vec = torch.cat([mean, meansqr], dim=0) vec = GroupGather.apply(vec, self.num_sync_devices, self.num_groups) * ( 1.0 / self.num_sync_devices ) mean, meansqr = torch.split(vec, C) var = meansqr - mean * mean self.running_mean += self.momentum * (mean.detach() - self.running_mean) self.running_var += self.momentum * (var.detach() - self.running_var) invstd = torch.rsqrt(var + self.eps) scale = self.weight * invstd bias = self.bias - mean * scale scale = scale.reshape(1, -1, 1, 1, 1) bias = bias.reshape(1, -1, 1, 1, 1) return input * scale + bias
7,424
32.90411
82
py
STTS
STTS-main/MViT/slowfast/models/mvit.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Video models.""" import math from functools import partial import torch import torch.nn as nn from torch.nn.init import trunc_normal_ from einops import rearrange from math import sqrt import slowfast.utils.weight_init_helper as init_helper from slowfast.models.attention import MultiScaleBlock from slowfast.models.batchnorm_helper import get_norm from slowfast.models.stem_helper import PatchEmbed from slowfast.models.utils import round_width, validate_checkpoint_wrapper_import from slowfast.models.topk import PatchNet from . import head_helper from .build import MODEL_REGISTRY try: from fairscale.nn.checkpoint import checkpoint_wrapper except ImportError: checkpoint_wrapper = None @MODEL_REGISTRY.register() class MViT(nn.Module): """ Multiscale Vision Transformers Haoqi Fan, Bo Xiong, Karttikeya Mangalam, Yanghao Li, Zhicheng Yan, Jitendra Malik, Christoph Feichtenhofer https://arxiv.org/abs/2104.11227 """ def __init__(self, cfg): super().__init__() # Get parameters. assert cfg.DATA.TRAIN_CROP_SIZE == cfg.DATA.TEST_CROP_SIZE self.cfg = cfg pool_first = cfg.MVIT.POOL_FIRST # Prepare input. spatial_size = cfg.DATA.TRAIN_CROP_SIZE temporal_size = cfg.DATA.NUM_FRAMES in_chans = cfg.DATA.INPUT_CHANNEL_NUM[0] use_2d_patch = cfg.MVIT.PATCH_2D self.patch_stride = cfg.MVIT.PATCH_STRIDE if use_2d_patch: self.patch_stride = [1] + self.patch_stride # Prepare output. num_classes = cfg.MODEL.NUM_CLASSES embed_dim = cfg.MVIT.EMBED_DIM # Prepare backbone num_heads = cfg.MVIT.NUM_HEADS mlp_ratio = cfg.MVIT.MLP_RATIO qkv_bias = cfg.MVIT.QKV_BIAS self.drop_rate = cfg.MVIT.DROPOUT_RATE depth = cfg.MVIT.DEPTH drop_path_rate = cfg.MVIT.DROPPATH_RATE mode = cfg.MVIT.MODE self.cls_embed_on = cfg.MVIT.CLS_EMBED_ON self.sep_pos_embed = cfg.MVIT.SEP_POS_EMBED if cfg.MVIT.NORM == "layernorm": norm_layer = partial(nn.LayerNorm, eps=1e-6) else: raise NotImplementedError("Only supports layernorm.") self.num_classes = num_classes self.patch_embed = PatchEmbed( dim_in=in_chans, dim_out=embed_dim, kernel=cfg.MVIT.PATCH_KERNEL, stride=cfg.MVIT.PATCH_STRIDE, padding=cfg.MVIT.PATCH_PADDING, conv_2d=use_2d_patch, ) self.input_dims = [temporal_size, spatial_size, spatial_size] assert self.input_dims[1] == self.input_dims[2] self.patch_dims = [ self.input_dims[i] // self.patch_stride[i] for i in range(len(self.input_dims)) ] num_patches = math.prod(self.patch_dims) dpr = [ x.item() for x in torch.linspace(0, drop_path_rate, depth) ] # stochastic depth decay rule if self.cls_embed_on: self.cls_token = nn.Parameter(torch.zeros(1, 1, embed_dim)) pos_embed_dim = num_patches + 1 else: pos_embed_dim = num_patches if self.sep_pos_embed: self.pos_embed_spatial = nn.Parameter( torch.zeros( 1, self.patch_dims[1] * self.patch_dims[2], embed_dim ) ) self.pos_embed_temporal = nn.Parameter( torch.zeros(1, self.patch_dims[0], embed_dim) ) if self.cls_embed_on: self.pos_embed_class = nn.Parameter( torch.zeros(1, 1, embed_dim) ) else: self.pos_embed = nn.Parameter( torch.zeros(1, pos_embed_dim, embed_dim) ) if self.drop_rate > 0.0: self.pos_drop = nn.Dropout(p=self.drop_rate) self.time_pruning_loc = cfg.MVIT.TIME_PRUNING_LOC time_left_ratio = cfg.MVIT.TIME_LEFT_RATIO time_score = cfg.MVIT.TIME_SCORE self.space_pruning_loc = cfg.MVIT.SPACE_PRUNING_LOC space_left_ratio = cfg.MVIT.SPACE_LEFT_RATIO space_score = cfg.MVIT.SPACE_SCORE self.sigma_max = cfg.MVIT.SIGMA self.sigma = cfg.MVIT.SIGMA dim_mul, head_mul = torch.ones(depth + 1), torch.ones(depth + 1) for i in range(len(cfg.MVIT.DIM_MUL)): dim_mul[cfg.MVIT.DIM_MUL[i][0]] = cfg.MVIT.DIM_MUL[i][1] for i in range(len(cfg.MVIT.HEAD_MUL)): head_mul[cfg.MVIT.HEAD_MUL[i][0]] = cfg.MVIT.HEAD_MUL[i][1] pool_q = [[] for i in range(cfg.MVIT.DEPTH)] pool_kv = [[] for i in range(cfg.MVIT.DEPTH)] stride_q = [[] for i in range(cfg.MVIT.DEPTH)] stride_kv = [[] for i in range(cfg.MVIT.DEPTH)] for i in range(len(cfg.MVIT.POOL_Q_STRIDE)): stride_q[cfg.MVIT.POOL_Q_STRIDE[i][0]] = cfg.MVIT.POOL_Q_STRIDE[i][ 1: ] if cfg.MVIT.POOL_KVQ_KERNEL is not None: pool_q[cfg.MVIT.POOL_Q_STRIDE[i][0]] = cfg.MVIT.POOL_KVQ_KERNEL else: pool_q[cfg.MVIT.POOL_Q_STRIDE[i][0]] = [ s + 1 if s > 1 else s for s in cfg.MVIT.POOL_Q_STRIDE[i][1:] ] # If POOL_KV_STRIDE_ADAPTIVE is not None, initialize POOL_KV_STRIDE. if cfg.MVIT.POOL_KV_STRIDE_ADAPTIVE is not None: _stride_kv = cfg.MVIT.POOL_KV_STRIDE_ADAPTIVE cfg.MVIT.POOL_KV_STRIDE = [] for i in range(cfg.MVIT.DEPTH): if len(stride_q[i]) > 0: _stride_kv = [ max(_stride_kv[d] // stride_q[i][d], 1) for d in range(len(_stride_kv)) ] cfg.MVIT.POOL_KV_STRIDE.append([i] + _stride_kv) for i in range(len(cfg.MVIT.POOL_KV_STRIDE)): stride_kv[cfg.MVIT.POOL_KV_STRIDE[i][0]] = cfg.MVIT.POOL_KV_STRIDE[ i ][1:] if cfg.MVIT.POOL_KVQ_KERNEL is not None: pool_kv[ cfg.MVIT.POOL_KV_STRIDE[i][0] ] = cfg.MVIT.POOL_KVQ_KERNEL else: pool_kv[cfg.MVIT.POOL_KV_STRIDE[i][0]] = [ s + 1 if s > 1 else s for s in cfg.MVIT.POOL_KV_STRIDE[i][1:] ] self.norm_stem = norm_layer(embed_dim) if cfg.MVIT.NORM_STEM else None self.blocks = nn.ModuleList() if cfg.MODEL.ACT_CHECKPOINT: validate_checkpoint_wrapper_import(checkpoint_wrapper) embedding_temporal_size = temporal_size // 2 embedding_spatial_size = self.patch_dims[1] * self.patch_dims[2] time_score_predictor = nn.ModuleList() space_score_predictor = nn.ModuleList() s_count = 0 t_count = 0 for i in range(depth): num_heads = round_width(num_heads, head_mul[i]) embed_dim = round_width(embed_dim, dim_mul[i], divisor=num_heads) dim_out = round_width( embed_dim, dim_mul[i + 1], divisor=round_width(num_heads, head_mul[i + 1]), ) attention_block = MultiScaleBlock( dim=embed_dim, dim_out=dim_out, num_heads=num_heads, mlp_ratio=mlp_ratio, qkv_bias=qkv_bias, drop_rate=self.drop_rate, drop_path=dpr[i], norm_layer=norm_layer, kernel_q=pool_q[i] if len(pool_q) > i else [], kernel_kv=pool_kv[i] if len(pool_kv) > i else [], stride_q=stride_q[i] if len(stride_q) > i else [], stride_kv=stride_kv[i] if len(stride_kv) > i else [], mode=mode, has_cls_embed=self.cls_embed_on, pool_first=pool_first, ) if cfg.MODEL.ACT_CHECKPOINT: attention_block = checkpoint_wrapper(attention_block) self.blocks.append(attention_block) if len(stride_q[i]) > 0: embedding_spatial_size = (int(sqrt(embedding_spatial_size)) // stride_q[i][1]) ** 2 if self.time_pruning_loc is not None and i in self.time_pruning_loc: left_frames = int(embedding_temporal_size * time_left_ratio[t_count]) t_count += 1 patchnet = PatchNet(score=time_score, k=left_frames, in_channels = embed_dim) time_score_predictor.append(patchnet) embedding_temporal_size = left_frames if self.space_pruning_loc is not None and i in self.space_pruning_loc: left_patches = int(embedding_spatial_size * space_left_ratio[s_count]) s_count += 1 patchnet = PatchNet(score=space_score, k=left_patches, in_channels = embed_dim) space_score_predictor.append(patchnet) embedding_spatial_size = left_patches if len(time_score_predictor) > 0: self.time_score_predictor = time_score_predictor if len(space_score_predictor) > 0: self.space_score_predictor = space_score_predictor embed_dim = dim_out self.norm = norm_layer(embed_dim) self.head = head_helper.TransformerBasicHead( embed_dim, num_classes, dropout_rate=cfg.MODEL.DROPOUT_RATE, act_func=cfg.MODEL.HEAD_ACT, ) if self.sep_pos_embed: trunc_normal_(self.pos_embed_spatial, std=0.02) trunc_normal_(self.pos_embed_temporal, std=0.02) if self.cls_embed_on: trunc_normal_(self.pos_embed_class, std=0.02) else: trunc_normal_(self.pos_embed, std=0.02) if self.cls_embed_on: trunc_normal_(self.cls_token, std=0.02) self.apply(self._init_weights) def _init_weights(self, m): if isinstance(m, nn.Linear): nn.init.trunc_normal_(m.weight, std=0.02) if isinstance(m, nn.Linear) and m.bias is not None: nn.init.constant_(m.bias, 0) elif isinstance(m, nn.LayerNorm): nn.init.constant_(m.bias, 0) nn.init.constant_(m.weight, 1.0) @torch.jit.ignore def no_weight_decay(self): if self.cfg.MVIT.ZERO_DECAY_POS_CLS: if self.sep_pos_embed: if self.cls_embed_on: return { "pos_embed_spatial", "pos_embed_temporal", "pos_embed_class", "cls_token", } else: return { "pos_embed_spatial", "pos_embed_temporal", "pos_embed_class", } else: if self.cls_embed_on: return {"pos_embed", "cls_token"} else: return {"pos_embed"} else: return {} def update_sigma(self, cur_step, total_steps): process = cur_step / total_steps sigma_multiplier = 1 - process self.sigma = self.sigma_max * sigma_multiplier def forward(self, x): x = x[0] x = self.patch_embed(x) T = self.cfg.DATA.NUM_FRAMES // self.patch_stride[0] H = self.cfg.DATA.TRAIN_CROP_SIZE // self.patch_stride[1] W = self.cfg.DATA.TRAIN_CROP_SIZE // self.patch_stride[2] B, TN, C = x.shape N = TN // T if self.cls_embed_on: cls_tokens = self.cls_token.expand( B, -1, -1 ) # stole cls_tokens impl from Phil Wang, thanks x = torch.cat((cls_tokens, x), dim=1) if self.sep_pos_embed: pos_embed = self.pos_embed_spatial.repeat( 1, self.patch_dims[0], 1 ) + torch.repeat_interleave( self.pos_embed_temporal, self.patch_dims[1] * self.patch_dims[2], dim=1, ) if self.cls_embed_on: pos_embed = torch.cat([self.pos_embed_class, pos_embed], 1) x = x + pos_embed else: x = x + self.pos_embed if self.drop_rate: x = self.pos_drop(x) if self.norm_stem: x = self.norm_stem(x) thw = [T, H, W] t_count = 0 s_count = 0 for i, blk in enumerate(self.blocks): if hasattr(self, 'time_score_predictor') and i in self.time_pruning_loc: if self.cls_embed_on: cls_tokens, x = x[:, 0:1], x[:,1:] x = self.time_score_predictor[t_count](x, 'time', N, T, self.sigma) T = x.size(1) // N t_count += 1 if self.cls_embed_on: x = torch.cat((cls_tokens, x), dim=1) thw = [T, H, W] if hasattr(self, 'space_score_predictor') and i in self.space_pruning_loc: if self.cls_embed_on: cls_tokens, x = x[:, 0:1, :], x[:,1:] x = self.space_score_predictor[s_count](x, 'space', N, T, self.sigma) N = x.size(1) // T H = W = int(math.sqrt(N)) s_count += 1 if self.cls_embed_on: x = torch.cat((cls_tokens, x), dim=1) thw = [T, H, W] x, thw = blk(x, thw) T, H, W = thw[0], thw[1], thw[1] N = H * W x = self.norm(x) if self.cls_embed_on: x = x[:, 0] else: x = x.mean(1) x = self.head(x) return x
14,098
35.058824
111
py
STTS
STTS-main/MViT/slowfast/models/ptv_model_builder.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Video models using PyTorchVideo model builder.""" from functools import partial import torch.nn as nn from detectron2.layers import ROIAlign from slowfast.models.batchnorm_helper import get_norm from slowfast.models.video_model_builder import _POOL1, _TEMPORAL_KERNEL_BASIS from pytorchvideo.models.csn import create_csn from pytorchvideo.models.head import ( create_res_basic_head, create_res_roi_pooling_head, ) from pytorchvideo.models.r2plus1d import ( create_2plus1d_bottleneck_block, create_r2plus1d, ) from pytorchvideo.models.resnet import create_bottleneck_block, create_resnet from pytorchvideo.models.slowfast import create_slowfast from pytorchvideo.models.x3d import ( Swish, create_x3d, create_x3d_bottleneck_block, ) from pytorchvideo.models.vision_transformers import create_multiscale_vision_transformers from .build import MODEL_REGISTRY def get_head_act(act_func): """ Return the actual head activation function given the activation fucntion name. Args: act_func (string): activation function to use. 'softmax': applies softmax on the output. 'sigmoid': applies sigmoid on the output. Returns: nn.Module: the activation layer. """ if act_func == "softmax": return nn.Softmax(dim=1) elif act_func == "sigmoid": return nn.Sigmoid() else: raise NotImplementedError( "{} is not supported as a head activation " "function.".format(act_func) ) @MODEL_REGISTRY.register() class PTVResNet(nn.Module): """ ResNet models using PyTorchVideo model builder. """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(PTVResNet, self).__init__() assert ( cfg.RESNET.STRIDE_1X1 is False ), "STRIDE_1x1 must be True for PTVResNet" assert ( cfg.RESNET.TRANS_FUNC == "bottleneck_transform" ), f"Unsupported TRANS_FUNC type {cfg.RESNET.TRANS_FUNC} for PTVResNet" assert cfg.MODEL.ARCH in [ "c2d", "slow", "i3d", ], f"Unsupported MODEL.ARCH type {cfg.MODEL.ARCH} for PTVResNet" self.detection_mode = cfg.DETECTION.ENABLE self._construct_network(cfg) def _construct_network(self, cfg): """ Builds a single pathway ResNet model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ # Params from configs. norm_module = get_norm(cfg) head_act = get_head_act(cfg.MODEL.HEAD_ACT) pool_size = _POOL1[cfg.MODEL.ARCH] num_groups = cfg.RESNET.NUM_GROUPS spatial_dilations = cfg.RESNET.SPATIAL_DILATIONS spatial_strides = cfg.RESNET.SPATIAL_STRIDES temp_kernel = _TEMPORAL_KERNEL_BASIS[cfg.MODEL.ARCH] stage1_pool = pool_size[0][0] != 1 or len(set(pool_size[0])) > 1 stage_spatial_stride = ( spatial_strides[0][0], spatial_strides[1][0], spatial_strides[2][0], spatial_strides[3][0], ) if cfg.MODEL.ARCH == "i3d": stage_conv_a_kernel_size = ( (3, 1, 1), [(3, 1, 1), (1, 1, 1)], [(3, 1, 1), (1, 1, 1)], [(1, 1, 1), (3, 1, 1)], ) else: stage_conv_a_kernel_size = ( (temp_kernel[1][0][0], 1, 1), (temp_kernel[2][0][0], 1, 1), (temp_kernel[3][0][0], 1, 1), (temp_kernel[4][0][0], 1, 1), ) # Head from config if cfg.DETECTION.ENABLE: self.detection_head = create_res_roi_pooling_head( in_features=cfg.RESNET.WIDTH_PER_GROUP * 2 ** (4 + 1), out_features=cfg.MODEL.NUM_CLASSES, pool=nn.AvgPool3d, output_size=(1, 1, 1), pool_kernel_size=( cfg.DATA.NUM_FRAMES // pool_size[0][0], 1, 1, ), dropout_rate=cfg.MODEL.DROPOUT_RATE, activation=None, output_with_global_average=False, pool_spatial=nn.MaxPool2d, resolution=[cfg.DETECTION.ROI_XFORM_RESOLUTION] * 2, spatial_scale=1.0 / float(cfg.DETECTION.SPATIAL_SCALE_FACTOR), sampling_ratio=0, roi=ROIAlign, ) self.model = create_resnet( # Input clip configs. input_channel=cfg.DATA.INPUT_CHANNEL_NUM[0], # Model configs. model_depth=cfg.RESNET.DEPTH, model_num_class=cfg.MODEL.NUM_CLASSES, dropout_rate=cfg.MODEL.DROPOUT_RATE, # Normalization configs. norm=norm_module, # Activation configs. activation=partial(nn.ReLU, inplace=cfg.RESNET.INPLACE_RELU), # Stem configs. stem_dim_out=cfg.RESNET.WIDTH_PER_GROUP, stem_conv_kernel_size=(temp_kernel[0][0][0], 7, 7), stem_conv_stride=(1, 2, 2), stem_pool=nn.MaxPool3d, stem_pool_kernel_size=(1, 3, 3), stem_pool_stride=(1, 2, 2), # Stage configs. stage1_pool=nn.MaxPool3d if stage1_pool else None, stage1_pool_kernel_size=pool_size[0], stage_conv_a_kernel_size=stage_conv_a_kernel_size, stage_conv_b_kernel_size=( (1, 3, 3), (1, 3, 3), (1, 3, 3), (1, 3, 3), ), stage_conv_b_num_groups=( num_groups, num_groups, num_groups, num_groups, ), stage_conv_b_dilation=( (1, spatial_dilations[0][0], spatial_dilations[0][0]), (1, spatial_dilations[1][0], spatial_dilations[1][0]), (1, spatial_dilations[2][0], spatial_dilations[2][0]), (1, spatial_dilations[3][0], spatial_dilations[3][0]), ), stage_spatial_h_stride=stage_spatial_stride, stage_spatial_w_stride=stage_spatial_stride, stage_temporal_stride=(1, 1, 1, 1), bottleneck=create_bottleneck_block, # Head configs. head=create_res_basic_head if not self.detection_mode else None, head_pool=nn.AvgPool3d, head_pool_kernel_size=( cfg.DATA.NUM_FRAMES // pool_size[0][0], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][1], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][2], ), head_activation=None, head_output_with_global_average=False, ) self.post_act = head_act def forward(self, x, bboxes=None): x = x[0] x = self.model(x) if self.detection_mode: x = self.detection_head(x, bboxes) x = self.post_act(x) else: # Performs fully convlutional inference. if not self.training: x = self.post_act(x) x = x.mean([2, 3, 4]) x = x.view(x.shape[0], -1) return x @MODEL_REGISTRY.register() class PTVSlowFast(nn.Module): def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(PTVSlowFast, self).__init__() assert ( cfg.RESNET.STRIDE_1X1 is False ), "STRIDE_1x1 must be True for PTVSlowFast" assert ( cfg.RESNET.TRANS_FUNC == "bottleneck_transform" ), f"Unsupported TRANS_FUNC type {cfg.RESNET.TRANS_FUNC} for PTVSlowFast" self.detection_mode = cfg.DETECTION.ENABLE self._construct_network(cfg) def _construct_network(self, cfg): """ Builds a SlowFast model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ _MODEL_STAGE_DEPTH = {50: (3, 4, 6, 3), 101: (3, 4, 23, 3)} # Params from configs. norm_module = get_norm(cfg) pool_size = _POOL1[cfg.MODEL.ARCH] num_groups = cfg.RESNET.NUM_GROUPS width_per_group = cfg.RESNET.WIDTH_PER_GROUP spatial_dilations = cfg.RESNET.SPATIAL_DILATIONS spatial_strides = cfg.RESNET.SPATIAL_STRIDES temp_kernel = _TEMPORAL_KERNEL_BASIS[cfg.MODEL.ARCH] num_block_temp_kernel = cfg.RESNET.NUM_BLOCK_TEMP_KERNEL stage_depth = _MODEL_STAGE_DEPTH[cfg.RESNET.DEPTH] stage_conv_a_kernel_sizes = [[], []] for pathway in range(2): for stage in range(4): stage_conv_a_kernel_sizes[pathway].append( ((temp_kernel[stage + 1][pathway][0], 1, 1),) * num_block_temp_kernel[stage][pathway] + ((1, 1, 1),) * ( stage_depth[stage] - num_block_temp_kernel[stage][pathway] ) ) # Head from config # Number of stages = 4 stage_dim_in = cfg.RESNET.WIDTH_PER_GROUP * 2 ** (4 + 1) head_in_features = stage_dim_in + stage_dim_in // cfg.SLOWFAST.BETA_INV if cfg.DETECTION.ENABLE: self.detection_head = create_res_roi_pooling_head( in_features=head_in_features, out_features=cfg.MODEL.NUM_CLASSES, pool=None, output_size=(1, 1, 1), dropout_rate=cfg.MODEL.DROPOUT_RATE, activation=None, output_with_global_average=False, pool_spatial=nn.MaxPool2d, resolution=[cfg.DETECTION.ROI_XFORM_RESOLUTION] * 2, spatial_scale=1.0 / float(cfg.DETECTION.SPATIAL_SCALE_FACTOR), sampling_ratio=0, roi=ROIAlign, ) head_pool_kernel_sizes = ( ( cfg.DATA.NUM_FRAMES // cfg.SLOWFAST.ALPHA // pool_size[0][0], 1, 1, ), (cfg.DATA.NUM_FRAMES // pool_size[1][0], 1, 1), ) else: head_pool_kernel_sizes = ( ( cfg.DATA.NUM_FRAMES // cfg.SLOWFAST.ALPHA // pool_size[0][0], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][1], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][2], ), ( cfg.DATA.NUM_FRAMES // pool_size[1][0], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[1][1], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[1][2], ), ) self.model = create_slowfast( # SlowFast configs. slowfast_channel_reduction_ratio=cfg.SLOWFAST.BETA_INV, slowfast_conv_channel_fusion_ratio=cfg.SLOWFAST.FUSION_CONV_CHANNEL_RATIO, slowfast_fusion_conv_kernel_size=( cfg.SLOWFAST.FUSION_KERNEL_SZ, 1, 1, ), slowfast_fusion_conv_stride=(cfg.SLOWFAST.ALPHA, 1, 1), # Input clip configs. input_channels=cfg.DATA.INPUT_CHANNEL_NUM, # Model configs. model_depth=cfg.RESNET.DEPTH, model_num_class=cfg.MODEL.NUM_CLASSES, dropout_rate=cfg.MODEL.DROPOUT_RATE, # Normalization configs. norm=norm_module, # Activation configs. activation=partial(nn.ReLU, inplace=cfg.RESNET.INPLACE_RELU), # Stem configs. stem_dim_outs=( width_per_group, width_per_group // cfg.SLOWFAST.BETA_INV, ), stem_conv_kernel_sizes=( (temp_kernel[0][0][0], 7, 7), (temp_kernel[0][1][0], 7, 7), ), stem_conv_strides=((1, 2, 2), (1, 2, 2)), stem_pool=nn.MaxPool3d, stem_pool_kernel_sizes=((1, 3, 3), (1, 3, 3)), stem_pool_strides=((1, 2, 2), (1, 2, 2)), # Stage configs. stage_conv_a_kernel_sizes=stage_conv_a_kernel_sizes, stage_conv_b_kernel_sizes=( ((1, 3, 3), (1, 3, 3), (1, 3, 3), (1, 3, 3)), ((1, 3, 3), (1, 3, 3), (1, 3, 3), (1, 3, 3)), ), stage_conv_b_num_groups=( (num_groups, num_groups, num_groups, num_groups), (num_groups, num_groups, num_groups, num_groups), ), stage_conv_b_dilations=( ( (1, spatial_dilations[0][0], spatial_dilations[0][0]), (1, spatial_dilations[1][0], spatial_dilations[1][0]), (1, spatial_dilations[2][0], spatial_dilations[2][0]), (1, spatial_dilations[3][0], spatial_dilations[3][0]), ), ( (1, spatial_dilations[0][1], spatial_dilations[0][1]), (1, spatial_dilations[1][1], spatial_dilations[1][1]), (1, spatial_dilations[1][1], spatial_dilations[1][1]), (1, spatial_dilations[1][1], spatial_dilations[1][1]), ), ), stage_spatial_strides=( ( spatial_strides[0][0], spatial_strides[1][0], spatial_strides[2][0], spatial_strides[3][0], ), ( spatial_strides[0][1], spatial_strides[1][1], spatial_strides[2][1], spatial_strides[3][1], ), ), stage_temporal_strides=((1, 1, 1, 1), (1, 1, 1, 1)), bottleneck=create_bottleneck_block, # Head configs. head=create_res_basic_head if not self.detection_mode else None, head_pool=nn.AvgPool3d, head_pool_kernel_sizes=head_pool_kernel_sizes, head_activation=None, head_output_with_global_average=False, ) self.post_act = get_head_act(cfg.MODEL.HEAD_ACT) def forward(self, x, bboxes=None): x = self.model(x) if self.detection_mode: x = self.detection_head(x, bboxes) x = self.post_act(x) else: # Performs fully convlutional inference. if not self.training: x = self.post_act(x) x = x.mean([2, 3, 4]) x = x.view(x.shape[0], -1) return x @MODEL_REGISTRY.register() class PTVX3D(nn.Module): def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(PTVX3D, self).__init__() assert ( cfg.RESNET.STRIDE_1X1 is False ), "STRIDE_1x1 must be True for PTVX3D" assert ( cfg.RESNET.TRANS_FUNC == "x3d_transform" ), f"Unsupported TRANS_FUNC type {cfg.RESNET.TRANS_FUNC} for PTVX3D" assert ( cfg.DETECTION.ENABLE is False ), "Detection model is not supported for PTVX3D yet." self._construct_network(cfg) def _construct_network(self, cfg): """ Builds a X3D model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ # Params from configs. norm_module = get_norm(cfg) temp_kernel = _TEMPORAL_KERNEL_BASIS[cfg.MODEL.ARCH] self.model = create_x3d( # Input clip configs. input_channel=cfg.DATA.INPUT_CHANNEL_NUM[0], input_clip_length=cfg.DATA.NUM_FRAMES, input_crop_size=cfg.DATA.TRAIN_CROP_SIZE, # Model configs. model_num_class=cfg.MODEL.NUM_CLASSES, dropout_rate=cfg.MODEL.DROPOUT_RATE, width_factor=cfg.X3D.WIDTH_FACTOR, depth_factor=cfg.X3D.DEPTH_FACTOR, # Normalization configs. norm=norm_module, norm_eps=1e-5, norm_momentum=0.1, # Activation configs. activation=partial(nn.ReLU, inplace=cfg.RESNET.INPLACE_RELU), # Stem configs. stem_dim_in=cfg.X3D.DIM_C1, stem_conv_kernel_size=(temp_kernel[0][0][0], 3, 3), stem_conv_stride=(1, 2, 2), # Stage configs. stage_conv_kernel_size=( (temp_kernel[1][0][0], 3, 3), (temp_kernel[2][0][0], 3, 3), (temp_kernel[3][0][0], 3, 3), (temp_kernel[4][0][0], 3, 3), ), stage_spatial_stride=(2, 2, 2, 2), stage_temporal_stride=(1, 1, 1, 1), bottleneck=create_x3d_bottleneck_block, bottleneck_factor=cfg.X3D.BOTTLENECK_FACTOR, se_ratio=0.0625, inner_act=Swish, # Head configs. head_dim_out=cfg.X3D.DIM_C5, head_pool_act=partial(nn.ReLU, inplace=cfg.RESNET.INPLACE_RELU), head_bn_lin5_on=cfg.X3D.BN_LIN5, head_activation=None, head_output_with_global_average=False, ) self.post_act = get_head_act(cfg.MODEL.HEAD_ACT) def forward(self, x, bboxes=None): x = x[0] x = self.model(x) # Performs fully convlutional inference. if not self.training: x = self.post_act(x) x = x.mean([2, 3, 4]) x = x.reshape(x.shape[0], -1) return x @MODEL_REGISTRY.register() class PTVCSN(nn.Module): """ CSN models using PyTorchVideo model builder. """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(PTVCSN, self).__init__() assert ( cfg.DETECTION.ENABLE is False ), "Detection model is not supported for PTVCSN yet." self._construct_network(cfg) def _construct_network(self, cfg): """ Builds a single pathway ResNet model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ # Params from configs. norm_module = get_norm(cfg) self.model = create_csn( # Input clip configs. input_channel=cfg.DATA.INPUT_CHANNEL_NUM[0], # Model configs. model_depth=cfg.RESNET.DEPTH, model_num_class=cfg.MODEL.NUM_CLASSES, dropout_rate=cfg.MODEL.DROPOUT_RATE, # Normalization configs. norm=norm_module, # Activation configs. activation=partial(nn.ReLU, inplace=cfg.RESNET.INPLACE_RELU), # Stem configs. stem_dim_out=cfg.RESNET.WIDTH_PER_GROUP, stem_conv_kernel_size=(3, 7, 7), stem_conv_stride=(1, 2, 2), stem_pool=nn.MaxPool3d, stem_pool_kernel_size=(1, 3, 3), stem_pool_stride=(1, 2, 2), # Stage configs. stage_conv_a_kernel_size=(1, 1, 1), stage_conv_b_kernel_size=(3, 3, 3), stage_conv_b_width_per_group=1, stage_spatial_stride=(1, 2, 2, 2), stage_temporal_stride=(1, 2, 2, 2), bottleneck=create_bottleneck_block, # Head configs. head_pool=nn.AvgPool3d, head_pool_kernel_size=( cfg.DATA.NUM_FRAMES // 8, cfg.DATA.TRAIN_CROP_SIZE // 32, cfg.DATA.TRAIN_CROP_SIZE // 32, ), head_activation=None, head_output_with_global_average=False, ) self.post_act = get_head_act(cfg.MODEL.HEAD_ACT) def forward(self, x, bboxes=None): x = x[0] x = self.model(x) # Performs fully convlutional inference. if not self.training: x = self.post_act(x) x = x.mean([2, 3, 4]) x = x.reshape(x.shape[0], -1) return x @MODEL_REGISTRY.register() class PTVR2plus1D(nn.Module): """ R(2+1)D models using PyTorchVideo model builder. """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(PTVR2plus1D, self).__init__() assert ( cfg.DETECTION.ENABLE is False ), "Detection model is not supported for PTVR2plus1D yet." self._construct_network(cfg) def _construct_network(self, cfg): """ Builds a single pathway R(2+1)D model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ self.model = create_r2plus1d( # Input clip configs. input_channel=cfg.DATA.INPUT_CHANNEL_NUM[0], # Model configs. model_depth=cfg.RESNET.DEPTH, model_num_class=cfg.MODEL.NUM_CLASSES, dropout_rate=cfg.MODEL.DROPOUT_RATE, # Normalization configs. norm=get_norm(cfg), norm_eps=1e-5, norm_momentum=0.1, # Activation configs. activation=partial(nn.ReLU, inplace=cfg.RESNET.INPLACE_RELU), # Stem configs. stem_dim_out=cfg.RESNET.WIDTH_PER_GROUP, stem_conv_kernel_size=(1, 7, 7), stem_conv_stride=(1, 2, 2), # Stage configs. stage_conv_a_kernel_size=( (1, 1, 1), (1, 1, 1), (1, 1, 1), (1, 1, 1), ), stage_conv_b_kernel_size=( (3, 3, 3), (3, 3, 3), (3, 3, 3), (3, 3, 3), ), stage_conv_b_num_groups=(1, 1, 1, 1), stage_conv_b_dilation=( (1, 1, 1), (1, 1, 1), (1, 1, 1), (1, 1, 1), ), stage_spatial_stride=(2, 2, 2, 2), stage_temporal_stride=(1, 1, 2, 2), stage_bottleneck=( create_2plus1d_bottleneck_block, create_2plus1d_bottleneck_block, create_2plus1d_bottleneck_block, create_2plus1d_bottleneck_block, ), # Head configs. head_pool=nn.AvgPool3d, head_pool_kernel_size=( cfg.DATA.NUM_FRAMES // 4, cfg.DATA.TRAIN_CROP_SIZE // 32, cfg.DATA.TRAIN_CROP_SIZE // 32, ), head_activation=None, head_output_with_global_average=False, ) self.post_act = get_head_act(cfg.MODEL.HEAD_ACT) def forward(self, x, bboxes=None): x = x[0] x = self.model(x) # Performs fully convlutional inference. if not self.training: x = self.post_act(x) x = x.mean([2, 3, 4]) x = x.view(x.shape[0], -1) return x @MODEL_REGISTRY.register() class PTVMViT(nn.Module): """ MViT models using PyTorchVideo model builder. """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(PTVMViT, self).__init__() assert ( cfg.DETECTION.ENABLE is False ), "Detection model is not supported for PTVMViT yet." self._construct_network(cfg) def _construct_network(self, cfg): """ Builds a MViT model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ self.model = create_multiscale_vision_transformers( spatial_size=cfg.DATA.TRAIN_CROP_SIZE, temporal_size=cfg.DATA.NUM_FRAMES, cls_embed_on=cfg.MVIT.CLS_EMBED_ON, sep_pos_embed=cfg.MVIT.SEP_POS_EMBED, depth=cfg.MVIT.DEPTH, norm=cfg.MVIT.NORM, # Patch embed config. input_channels = cfg.DATA.INPUT_CHANNEL_NUM[0], patch_embed_dim = cfg.MVIT.EMBED_DIM, conv_patch_embed_kernel = cfg.MVIT.PATCH_KERNEL, conv_patch_embed_stride = cfg.MVIT.PATCH_STRIDE, conv_patch_embed_padding = cfg.MVIT.PATCH_PADDING, enable_patch_embed_norm = cfg.MVIT.NORM_STEM, use_2d_patch=cfg.MVIT.PATCH_2D, # Attention block config. num_heads = cfg.MVIT.NUM_HEADS, mlp_ratio = cfg.MVIT.MLP_RATIO, qkv_bias = cfg.MVIT.QKV_BIAS, dropout_rate_block = cfg.MVIT.DROPOUT_RATE, droppath_rate_block = cfg.MVIT.DROPPATH_RATE, pooling_mode = cfg.MVIT.MODE, pool_first = cfg.MVIT.POOL_FIRST, embed_dim_mul = cfg.MVIT.DIM_MUL, atten_head_mul = cfg.MVIT.HEAD_MUL, pool_q_stride_size = cfg.MVIT.POOL_Q_STRIDE, pool_kv_stride_size = cfg.MVIT.POOL_KV_STRIDE, pool_kv_stride_adaptive = cfg.MVIT.POOL_KV_STRIDE_ADAPTIVE, pool_kvq_kernel = cfg.MVIT.POOL_KVQ_KERNEL, # Head config. head_dropout_rate = cfg.MODEL.DROPOUT_RATE, head_num_classes = cfg.MODEL.NUM_CLASSES, ) self.post_act = get_head_act(cfg.MODEL.HEAD_ACT) def forward(self, x, bboxes=None): x = x[0] x = self.model(x) if not self.training: x = self.post_act(x) return x
26,986
33.777062
89
py
STTS
STTS-main/MViT/slowfast/models/topk.py
import torch import torch.nn as nn import torch.nn.functional as F import numpy as np import einops from einops import rearrange from math import sqrt import time class PredictorLG(nn.Module): """ Image to Patch Embedding """ def __init__(self, embed_dim=384): super().__init__() self.in_conv = nn.Sequential( nn.LayerNorm(embed_dim), nn.Linear(embed_dim, embed_dim), nn.GELU() ) self.out_conv = nn.Sequential( nn.Linear(embed_dim, embed_dim // 2), nn.GELU(), nn.Linear(embed_dim // 2, embed_dim // 4), nn.GELU(), nn.Linear(embed_dim // 4, 1) ) def forward(self, x): x = self.in_conv(x) B, N, C = x.size() local_x = x[:,:, :C//2] global_x = torch.mean(x[:,:, C//2:], dim=1, keepdim=True) x = torch.cat([local_x, global_x.expand(B, N, C//2)], dim=-1) return self.out_conv(x) def HardTopK(k, x): topk_results = torch.topk(x, k=k, dim=-1, sorted=False) indices = topk_results.indices # b, k indices = torch.sort(indices, dim=-1).values return indices class PerturbedTopK(nn.Module): def __init__(self, k: int, num_samples: int = 1000): super(PerturbedTopK, self).__init__() self.num_samples = num_samples self.k = k def __call__(self, x, sigma): return PerturbedTopKFunction.apply(x, self.k, self.num_samples, sigma) class PerturbedTopKFunction(torch.autograd.Function): @staticmethod def forward(ctx, x, k: int, num_samples: int = 1000, sigma: float = 0.05): b, d = x.shape # for Gaussian: noise and gradient are the same. noise = torch.normal(mean=0.0, std=1.0, size=(b, num_samples, d)).to(x.device) perturbed_x = x[:, None, :] + noise * sigma # b, nS, d topk_results = torch.topk(perturbed_x, k=k, dim=-1, sorted=False) indices = topk_results.indices # b, nS, k indices = torch.sort(indices, dim=-1).values # b, nS, k perturbed_output = torch.nn.functional.one_hot(indices, num_classes=d).float() indicators = perturbed_output.mean(dim=1) # b, k, d # constants for backward ctx.k = k ctx.num_samples = num_samples ctx.sigma = sigma # tensors for backward ctx.perturbed_output = perturbed_output ctx.noise = noise return indicators @staticmethod def backward(ctx, grad_output): if grad_output is None: return tuple([None] * 5) noise_gradient = ctx.noise if ctx.sigma <= 1e-20: b, _, k, d = ctx.perturbed_output.size() expected_gradient = torch.zeros(b, k, d).to(grad_output.device) else: expected_gradient = ( torch.einsum("bnkd,bnd->bkd", ctx.perturbed_output, noise_gradient) / ctx.num_samples / (ctx.sigma) ) grad_input = torch.einsum("bkd,bkd->bd", grad_output, expected_gradient) return (grad_input,) + tuple([None] * 5) def batched_index_select(input, dim, index): for i in range(1, len(input.shape)): if i != dim: index = index.unsqueeze(i) expanse = list(input.shape) expanse[0] = -1 expanse[dim] = -1 index = index.expand(expanse) return torch.gather(input, dim, index) def extract_patches_from_indices(x, indices): batch_size, _, channels = x.shape k = indices.shape[-1] patches = x patches = batched_index_select(patches, 1, indices) patches = patches.contiguous().view(batch_size, k, channels) return patches def extract_patches_from_indicators(x, indicators): indicators = rearrange(indicators, "b d k -> b k d") patches = torch.einsum("b k d, b d c -> b k c", indicators, x) return patches def min_max_norm(x): flatten_score_min = x.min(axis=-1, keepdim=True).values flatten_score_max = x.max(axis=-1, keepdim=True).values norm_flatten_score = (x - flatten_score_min) / (flatten_score_max - flatten_score_min + 1e-5) return norm_flatten_score class PatchNet(nn.Module): def __init__(self, score, k, in_channels, stride=None, num_samples=500): super(PatchNet, self).__init__() self.k = k self.anchor_size = int(sqrt(k)) self.stride = stride self.score = score self.in_channels = in_channels self.num_samples = num_samples if score == 'tpool': self.score_network = PredictorLG(embed_dim=2*in_channels) elif score == 'spatch': self.score_network = PredictorLG(embed_dim=in_channels) self.init = torch.eye(self.k).unsqueeze(0).unsqueeze(-1).cuda() def get_indicator(self, scores, k, sigma): indicator = PerturbedTopKFunction.apply(scores, k, self.num_samples, sigma) indicator = einops.rearrange(indicator, "b k d -> b d k") return indicator def get_indices(self, scores, k): indices = HardTopK(k, scores) return indices def generate_random_indices(self, b, n, k): indices = [] for _ in range(b): indice = np.sort(np.random.choice(n, k, replace=False)) indices.append(indice) indices = np.vstack(indices) indices = torch.Tensor(indices).long().cuda() return indices def generate_uniform_indices(self, b, n, k): indices = torch.linspace(0, n-1, steps=k).long() indices = indices.unsqueeze(0).cuda() indices = indices.repeat(b, 1) return indices def forward(self, x, type, N, T, sigma): B = x.size(0) H = W = int(sqrt(N)) indicator = None indices = None if type == 'time': if self.score == 'tpool': x = rearrange(x, 'b (t n) m -> b t n m', t=T) avg = torch.mean(x, dim=2, keepdim=False) max_ = torch.max(x, dim=2).values x_ = torch.cat((avg, max_), dim=2) scores = self.score_network(x_).squeeze(-1) scores = min_max_norm(scores) if self.training: indicator = self.get_indicator(scores, self.k, sigma) else: indices = self.get_indices(scores, self.k) x = rearrange(x, 'b t n m -> b t (n m)') else: s = self.stride if self.stride is not None else int(max((H - self.anchor_size) // 2, 1)) if self.score == 'spatch': x = rearrange(x, 'b (t n) c -> (b t) n c', t=T) scores = self.score_network(x) scores = rearrange(scores, '(b t) (h w) c -> (b t) c h w', b=B, h=H) scores = F.unfold(scores, kernel_size=self.anchor_size, stride=s) scores = scores.mean(dim=1) scores = min_max_norm(scores) x = rearrange(x, '(b t) (h w) c -> (b t) c h w', b=B, h=H) x = F.unfold(x, kernel_size=self.anchor_size, stride=s).permute(0, 2, 1).contiguous() if self.training: indicator = self.get_indicator(scores, 1, sigma) else: indices = self.get_indices(scores, 1) if self.training: if indicator is not None: patches = extract_patches_from_indicators(x, indicator) elif indices is not None: patches = extract_patches_from_indices(x, indices) if type == 'time': patches = rearrange(patches, 'b k (n c) -> b (k n) c', n = N) elif self.score == 'spatch': patches = rearrange(patches, '(b t) k (c kh kw) -> b (t k kh kw) c', b=B, c=self.in_channels, kh=self.anchor_size) return patches else: patches = extract_patches_from_indices(x, indices) if type == 'time': patches = rearrange(patches, 'b k (n c) -> b (k n) c', n = N) elif self.score == 'spatch': patches = rearrange(patches, '(b t) k (c kh kw) -> b (t k kh kw) c', b=B, c=self.in_channels, kh=self.anchor_size) return patches
8,512
32.916335
101
py
STTS
STTS-main/MViT/slowfast/models/utils.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. from logging import NOTSET import slowfast.utils.logging as logging import numpy as np import torch.nn.functional as F from einops import rearrange logger = logging.get_logger(__name__) def round_width(width, multiplier, min_width=1, divisor=1, verbose=False): if not multiplier: return width width *= multiplier min_width = min_width or divisor if verbose: logger.info(f"min width {min_width}") logger.info(f"width {width} divisor {divisor}") logger.info(f"other {int(width + divisor / 2) // divisor * divisor}") width_out = max(min_width, int(width + divisor / 2) // divisor * divisor) if width_out < 0.9 * width: width_out += divisor return int(width_out) def validate_checkpoint_wrapper_import(checkpoint_wrapper): """ Check if checkpoint_wrapper is imported. """ if checkpoint_wrapper is None: raise ImportError("Please install fairscale.")
1,017
28.941176
77
py
STTS
STTS-main/MViT/slowfast/models/nonlocal_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Non-local helper""" import torch import torch.nn as nn class Nonlocal(nn.Module): """ Builds Non-local Neural Networks as a generic family of building blocks for capturing long-range dependencies. Non-local Network computes the response at a position as a weighted sum of the features at all positions. This building block can be plugged into many computer vision architectures. More details in the paper: https://arxiv.org/pdf/1711.07971.pdf """ def __init__( self, dim, dim_inner, pool_size=None, instantiation="softmax", zero_init_final_conv=False, zero_init_final_norm=True, norm_eps=1e-5, norm_momentum=0.1, norm_module=nn.BatchNorm3d, ): """ Args: dim (int): number of dimension for the input. dim_inner (int): number of dimension inside of the Non-local block. pool_size (list): the kernel size of spatial temporal pooling, temporal pool kernel size, spatial pool kernel size, spatial pool kernel size in order. By default pool_size is None, then there would be no pooling used. instantiation (string): supports two different instantiation method: "dot_product": normalizing correlation matrix with L2. "softmax": normalizing correlation matrix with Softmax. zero_init_final_conv (bool): If true, zero initializing the final convolution of the Non-local block. zero_init_final_norm (bool): If true, zero initializing the final batch norm of the Non-local block. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. """ super(Nonlocal, self).__init__() self.dim = dim self.dim_inner = dim_inner self.pool_size = pool_size self.instantiation = instantiation self.use_pool = ( False if pool_size is None else any((size > 1 for size in pool_size)) ) self.norm_eps = norm_eps self.norm_momentum = norm_momentum self._construct_nonlocal( zero_init_final_conv, zero_init_final_norm, norm_module ) def _construct_nonlocal( self, zero_init_final_conv, zero_init_final_norm, norm_module ): # Three convolution heads: theta, phi, and g. self.conv_theta = nn.Conv3d( self.dim, self.dim_inner, kernel_size=1, stride=1, padding=0 ) self.conv_phi = nn.Conv3d( self.dim, self.dim_inner, kernel_size=1, stride=1, padding=0 ) self.conv_g = nn.Conv3d( self.dim, self.dim_inner, kernel_size=1, stride=1, padding=0 ) # Final convolution output. self.conv_out = nn.Conv3d( self.dim_inner, self.dim, kernel_size=1, stride=1, padding=0 ) # Zero initializing the final convolution output. self.conv_out.zero_init = zero_init_final_conv # TODO: change the name to `norm` self.bn = norm_module( num_features=self.dim, eps=self.norm_eps, momentum=self.norm_momentum, ) # Zero initializing the final bn. self.bn.transform_final_bn = zero_init_final_norm # Optional to add the spatial-temporal pooling. if self.use_pool: self.pool = nn.MaxPool3d( kernel_size=self.pool_size, stride=self.pool_size, padding=[0, 0, 0], ) def forward(self, x): x_identity = x N, C, T, H, W = x.size() theta = self.conv_theta(x) # Perform temporal-spatial pooling to reduce the computation. if self.use_pool: x = self.pool(x) phi = self.conv_phi(x) g = self.conv_g(x) theta = theta.view(N, self.dim_inner, -1) phi = phi.view(N, self.dim_inner, -1) g = g.view(N, self.dim_inner, -1) # (N, C, TxHxW) * (N, C, TxHxW) => (N, TxHxW, TxHxW). theta_phi = torch.einsum("nct,ncp->ntp", (theta, phi)) # For original Non-local paper, there are two main ways to normalize # the affinity tensor: # 1) Softmax normalization (norm on exp). # 2) dot_product normalization. if self.instantiation == "softmax": # Normalizing the affinity tensor theta_phi before softmax. theta_phi = theta_phi * (self.dim_inner ** -0.5) theta_phi = nn.functional.softmax(theta_phi, dim=2) elif self.instantiation == "dot_product": spatial_temporal_dim = theta_phi.shape[2] theta_phi = theta_phi / spatial_temporal_dim else: raise NotImplementedError( "Unknown norm type {}".format(self.instantiation) ) # (N, TxHxW, TxHxW) * (N, C, TxHxW) => (N, C, TxHxW). theta_phi_g = torch.einsum("ntg,ncg->nct", (theta_phi, g)) # (N, C, TxHxW) => (N, C, T, H, W). theta_phi_g = theta_phi_g.view(N, self.dim_inner, T, H, W) p = self.conv_out(theta_phi_g) p = self.bn(p) return x_identity + p
5,418
35.369128
80
py
STTS
STTS-main/MViT/slowfast/models/video_model_builder.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Video models.""" import math from functools import partial import torch import torch.nn as nn from torch.nn.init import trunc_normal_ import slowfast.utils.weight_init_helper as init_helper from slowfast.models.attention import MultiScaleBlock from slowfast.models.batchnorm_helper import get_norm from slowfast.models.stem_helper import PatchEmbed from slowfast.models.utils import round_width, validate_checkpoint_wrapper_import from . import head_helper, resnet_helper, stem_helper from .build import MODEL_REGISTRY try: from fairscale.nn.checkpoint import checkpoint_wrapper except ImportError: checkpoint_wrapper = None # Number of blocks for different stages given the model depth. _MODEL_STAGE_DEPTH = {50: (3, 4, 6, 3), 101: (3, 4, 23, 3)} # Basis of temporal kernel sizes for each of the stage. _TEMPORAL_KERNEL_BASIS = { "2d": [ [[1]], # conv1 temporal kernel. [[1]], # res2 temporal kernel. [[1]], # res3 temporal kernel. [[1]], # res4 temporal kernel. [[1]], # res5 temporal kernel. ], "c2d": [ [[1]], # conv1 temporal kernel. [[1]], # res2 temporal kernel. [[1]], # res3 temporal kernel. [[1]], # res4 temporal kernel. [[1]], # res5 temporal kernel. ], "c2d_nopool": [ [[1]], # conv1 temporal kernel. [[1]], # res2 temporal kernel. [[1]], # res3 temporal kernel. [[1]], # res4 temporal kernel. [[1]], # res5 temporal kernel. ], "i3d": [ [[5]], # conv1 temporal kernel. [[3]], # res2 temporal kernel. [[3, 1]], # res3 temporal kernel. [[3, 1]], # res4 temporal kernel. [[1, 3]], # res5 temporal kernel. ], "i3d_nopool": [ [[5]], # conv1 temporal kernel. [[3]], # res2 temporal kernel. [[3, 1]], # res3 temporal kernel. [[3, 1]], # res4 temporal kernel. [[1, 3]], # res5 temporal kernel. ], "slow": [ [[1]], # conv1 temporal kernel. [[1]], # res2 temporal kernel. [[1]], # res3 temporal kernel. [[3]], # res4 temporal kernel. [[3]], # res5 temporal kernel. ], "slowfast": [ [[1], [5]], # conv1 temporal kernel for slow and fast pathway. [[1], [3]], # res2 temporal kernel for slow and fast pathway. [[1], [3]], # res3 temporal kernel for slow and fast pathway. [[3], [3]], # res4 temporal kernel for slow and fast pathway. [[3], [3]], # res5 temporal kernel for slow and fast pathway. ], "x3d": [ [[5]], # conv1 temporal kernels. [[3]], # res2 temporal kernels. [[3]], # res3 temporal kernels. [[3]], # res4 temporal kernels. [[3]], # res5 temporal kernels. ], } _POOL1 = { "2d": [[1, 1, 1]], "c2d": [[2, 1, 1]], "c2d_nopool": [[1, 1, 1]], "i3d": [[2, 1, 1]], "i3d_nopool": [[1, 1, 1]], "slow": [[1, 1, 1]], "slowfast": [[1, 1, 1], [1, 1, 1]], "x3d": [[1, 1, 1]], } class FuseFastToSlow(nn.Module): """ Fuses the information from the Fast pathway to the Slow pathway. Given the tensors from Slow pathway and Fast pathway, fuse information from Fast to Slow, then return the fused tensors from Slow and Fast pathway in order. """ def __init__( self, dim_in, fusion_conv_channel_ratio, fusion_kernel, alpha, eps=1e-5, bn_mmt=0.1, inplace_relu=True, norm_module=nn.BatchNorm3d, ): """ Args: dim_in (int): the channel dimension of the input. fusion_conv_channel_ratio (int): channel ratio for the convolution used to fuse from Fast pathway to Slow pathway. fusion_kernel (int): kernel size of the convolution used to fuse from Fast pathway to Slow pathway. alpha (int): the frame rate ratio between the Fast and Slow pathway. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. inplace_relu (bool): if True, calculate the relu on the original input without allocating new memory. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. """ super(FuseFastToSlow, self).__init__() self.conv_f2s = nn.Conv3d( dim_in, dim_in * fusion_conv_channel_ratio, kernel_size=[fusion_kernel, 1, 1], stride=[alpha, 1, 1], padding=[fusion_kernel // 2, 0, 0], bias=False, ) self.bn = norm_module( num_features=dim_in * fusion_conv_channel_ratio, eps=eps, momentum=bn_mmt, ) self.relu = nn.ReLU(inplace_relu) def forward(self, x): x_s = x[0] x_f = x[1] fuse = self.conv_f2s(x_f) fuse = self.bn(fuse) fuse = self.relu(fuse) x_s_fuse = torch.cat([x_s, fuse], 1) return [x_s_fuse, x_f] @MODEL_REGISTRY.register() class SlowFast(nn.Module): """ SlowFast model builder for SlowFast network. Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. "SlowFast networks for video recognition." https://arxiv.org/pdf/1812.03982.pdf """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(SlowFast, self).__init__() self.norm_module = get_norm(cfg) self.enable_detection = cfg.DETECTION.ENABLE self.num_pathways = 2 self._construct_network(cfg) init_helper.init_weights( self, cfg.MODEL.FC_INIT_STD, cfg.RESNET.ZERO_INIT_FINAL_BN ) def _construct_network(self, cfg): """ Builds a SlowFast model. The first pathway is the Slow pathway and the second pathway is the Fast pathway. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ assert cfg.MODEL.ARCH in _POOL1.keys() pool_size = _POOL1[cfg.MODEL.ARCH] assert len({len(pool_size), self.num_pathways}) == 1 assert cfg.RESNET.DEPTH in _MODEL_STAGE_DEPTH.keys() (d2, d3, d4, d5) = _MODEL_STAGE_DEPTH[cfg.RESNET.DEPTH] num_groups = cfg.RESNET.NUM_GROUPS width_per_group = cfg.RESNET.WIDTH_PER_GROUP dim_inner = num_groups * width_per_group out_dim_ratio = ( cfg.SLOWFAST.BETA_INV // cfg.SLOWFAST.FUSION_CONV_CHANNEL_RATIO ) temp_kernel = _TEMPORAL_KERNEL_BASIS[cfg.MODEL.ARCH] self.s1 = stem_helper.VideoModelStem( dim_in=cfg.DATA.INPUT_CHANNEL_NUM, dim_out=[width_per_group, width_per_group // cfg.SLOWFAST.BETA_INV], kernel=[temp_kernel[0][0] + [7, 7], temp_kernel[0][1] + [7, 7]], stride=[[1, 2, 2]] * 2, padding=[ [temp_kernel[0][0][0] // 2, 3, 3], [temp_kernel[0][1][0] // 2, 3, 3], ], norm_module=self.norm_module, ) self.s1_fuse = FuseFastToSlow( width_per_group // cfg.SLOWFAST.BETA_INV, cfg.SLOWFAST.FUSION_CONV_CHANNEL_RATIO, cfg.SLOWFAST.FUSION_KERNEL_SZ, cfg.SLOWFAST.ALPHA, norm_module=self.norm_module, ) self.s2 = resnet_helper.ResStage( dim_in=[ width_per_group + width_per_group // out_dim_ratio, width_per_group // cfg.SLOWFAST.BETA_INV, ], dim_out=[ width_per_group * 4, width_per_group * 4 // cfg.SLOWFAST.BETA_INV, ], dim_inner=[dim_inner, dim_inner // cfg.SLOWFAST.BETA_INV], temp_kernel_sizes=temp_kernel[1], stride=cfg.RESNET.SPATIAL_STRIDES[0], num_blocks=[d2] * 2, num_groups=[num_groups] * 2, num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[0], nonlocal_inds=cfg.NONLOCAL.LOCATION[0], nonlocal_group=cfg.NONLOCAL.GROUP[0], nonlocal_pool=cfg.NONLOCAL.POOL[0], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, dilation=cfg.RESNET.SPATIAL_DILATIONS[0], norm_module=self.norm_module, ) self.s2_fuse = FuseFastToSlow( width_per_group * 4 // cfg.SLOWFAST.BETA_INV, cfg.SLOWFAST.FUSION_CONV_CHANNEL_RATIO, cfg.SLOWFAST.FUSION_KERNEL_SZ, cfg.SLOWFAST.ALPHA, norm_module=self.norm_module, ) for pathway in range(self.num_pathways): pool = nn.MaxPool3d( kernel_size=pool_size[pathway], stride=pool_size[pathway], padding=[0, 0, 0], ) self.add_module("pathway{}_pool".format(pathway), pool) self.s3 = resnet_helper.ResStage( dim_in=[ width_per_group * 4 + width_per_group * 4 // out_dim_ratio, width_per_group * 4 // cfg.SLOWFAST.BETA_INV, ], dim_out=[ width_per_group * 8, width_per_group * 8 // cfg.SLOWFAST.BETA_INV, ], dim_inner=[dim_inner * 2, dim_inner * 2 // cfg.SLOWFAST.BETA_INV], temp_kernel_sizes=temp_kernel[2], stride=cfg.RESNET.SPATIAL_STRIDES[1], num_blocks=[d3] * 2, num_groups=[num_groups] * 2, num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[1], nonlocal_inds=cfg.NONLOCAL.LOCATION[1], nonlocal_group=cfg.NONLOCAL.GROUP[1], nonlocal_pool=cfg.NONLOCAL.POOL[1], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, dilation=cfg.RESNET.SPATIAL_DILATIONS[1], norm_module=self.norm_module, ) self.s3_fuse = FuseFastToSlow( width_per_group * 8 // cfg.SLOWFAST.BETA_INV, cfg.SLOWFAST.FUSION_CONV_CHANNEL_RATIO, cfg.SLOWFAST.FUSION_KERNEL_SZ, cfg.SLOWFAST.ALPHA, norm_module=self.norm_module, ) self.s4 = resnet_helper.ResStage( dim_in=[ width_per_group * 8 + width_per_group * 8 // out_dim_ratio, width_per_group * 8 // cfg.SLOWFAST.BETA_INV, ], dim_out=[ width_per_group * 16, width_per_group * 16 // cfg.SLOWFAST.BETA_INV, ], dim_inner=[dim_inner * 4, dim_inner * 4 // cfg.SLOWFAST.BETA_INV], temp_kernel_sizes=temp_kernel[3], stride=cfg.RESNET.SPATIAL_STRIDES[2], num_blocks=[d4] * 2, num_groups=[num_groups] * 2, num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[2], nonlocal_inds=cfg.NONLOCAL.LOCATION[2], nonlocal_group=cfg.NONLOCAL.GROUP[2], nonlocal_pool=cfg.NONLOCAL.POOL[2], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, dilation=cfg.RESNET.SPATIAL_DILATIONS[2], norm_module=self.norm_module, ) self.s4_fuse = FuseFastToSlow( width_per_group * 16 // cfg.SLOWFAST.BETA_INV, cfg.SLOWFAST.FUSION_CONV_CHANNEL_RATIO, cfg.SLOWFAST.FUSION_KERNEL_SZ, cfg.SLOWFAST.ALPHA, norm_module=self.norm_module, ) self.s5 = resnet_helper.ResStage( dim_in=[ width_per_group * 16 + width_per_group * 16 // out_dim_ratio, width_per_group * 16 // cfg.SLOWFAST.BETA_INV, ], dim_out=[ width_per_group * 32, width_per_group * 32 // cfg.SLOWFAST.BETA_INV, ], dim_inner=[dim_inner * 8, dim_inner * 8 // cfg.SLOWFAST.BETA_INV], temp_kernel_sizes=temp_kernel[4], stride=cfg.RESNET.SPATIAL_STRIDES[3], num_blocks=[d5] * 2, num_groups=[num_groups] * 2, num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[3], nonlocal_inds=cfg.NONLOCAL.LOCATION[3], nonlocal_group=cfg.NONLOCAL.GROUP[3], nonlocal_pool=cfg.NONLOCAL.POOL[3], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, dilation=cfg.RESNET.SPATIAL_DILATIONS[3], norm_module=self.norm_module, ) if cfg.DETECTION.ENABLE: self.head = head_helper.ResNetRoIHead( dim_in=[ width_per_group * 32, width_per_group * 32 // cfg.SLOWFAST.BETA_INV, ], num_classes=cfg.MODEL.NUM_CLASSES, pool_size=[ [ cfg.DATA.NUM_FRAMES // cfg.SLOWFAST.ALPHA // pool_size[0][0], 1, 1, ], [cfg.DATA.NUM_FRAMES // pool_size[1][0], 1, 1], ], resolution=[[cfg.DETECTION.ROI_XFORM_RESOLUTION] * 2] * 2, scale_factor=[cfg.DETECTION.SPATIAL_SCALE_FACTOR] * 2, dropout_rate=cfg.MODEL.DROPOUT_RATE, act_func=cfg.MODEL.HEAD_ACT, aligned=cfg.DETECTION.ALIGNED, ) else: self.head = head_helper.ResNetBasicHead( dim_in=[ width_per_group * 32, width_per_group * 32 // cfg.SLOWFAST.BETA_INV, ], num_classes=cfg.MODEL.NUM_CLASSES, pool_size=[None, None] if cfg.MULTIGRID.SHORT_CYCLE else [ [ cfg.DATA.NUM_FRAMES // cfg.SLOWFAST.ALPHA // pool_size[0][0], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][1], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][2], ], [ cfg.DATA.NUM_FRAMES // pool_size[1][0], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[1][1], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[1][2], ], ], # None for AdaptiveAvgPool3d((1, 1, 1)) dropout_rate=cfg.MODEL.DROPOUT_RATE, act_func=cfg.MODEL.HEAD_ACT, ) def forward(self, x, bboxes=None): x = self.s1(x) x = self.s1_fuse(x) x = self.s2(x) x = self.s2_fuse(x) for pathway in range(self.num_pathways): pool = getattr(self, "pathway{}_pool".format(pathway)) x[pathway] = pool(x[pathway]) x = self.s3(x) x = self.s3_fuse(x) x = self.s4(x) x = self.s4_fuse(x) x = self.s5(x) if self.enable_detection: x = self.head(x, bboxes) else: x = self.head(x) return x @MODEL_REGISTRY.register() class ResNet(nn.Module): """ ResNet model builder. It builds a ResNet like network backbone without lateral connection (C2D, I3D, Slow). Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. "SlowFast networks for video recognition." https://arxiv.org/pdf/1812.03982.pdf Xiaolong Wang, Ross Girshick, Abhinav Gupta, and Kaiming He. "Non-local neural networks." https://arxiv.org/pdf/1711.07971.pdf """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(ResNet, self).__init__() self.norm_module = get_norm(cfg) self.enable_detection = cfg.DETECTION.ENABLE self.num_pathways = 1 self._construct_network(cfg) init_helper.init_weights( self, cfg.MODEL.FC_INIT_STD, cfg.RESNET.ZERO_INIT_FINAL_BN ) def _construct_network(self, cfg): """ Builds a single pathway ResNet model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ assert cfg.MODEL.ARCH in _POOL1.keys() pool_size = _POOL1[cfg.MODEL.ARCH] assert len({len(pool_size), self.num_pathways}) == 1 assert cfg.RESNET.DEPTH in _MODEL_STAGE_DEPTH.keys() (d2, d3, d4, d5) = _MODEL_STAGE_DEPTH[cfg.RESNET.DEPTH] num_groups = cfg.RESNET.NUM_GROUPS width_per_group = cfg.RESNET.WIDTH_PER_GROUP dim_inner = num_groups * width_per_group temp_kernel = _TEMPORAL_KERNEL_BASIS[cfg.MODEL.ARCH] s1 = stem_helper.VideoModelStem( dim_in=cfg.DATA.INPUT_CHANNEL_NUM, dim_out=[width_per_group], kernel=[temp_kernel[0][0] + [7, 7]], stride=[[1, 2, 2]], padding=[[temp_kernel[0][0][0] // 2, 3, 3]], norm_module=self.norm_module, ) s2 = resnet_helper.ResStage( dim_in=[width_per_group], dim_out=[width_per_group * 4], dim_inner=[dim_inner], temp_kernel_sizes=temp_kernel[1], stride=cfg.RESNET.SPATIAL_STRIDES[0], num_blocks=[d2], num_groups=[num_groups], num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[0], nonlocal_inds=cfg.NONLOCAL.LOCATION[0], nonlocal_group=cfg.NONLOCAL.GROUP[0], nonlocal_pool=cfg.NONLOCAL.POOL[0], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, stride_1x1=cfg.RESNET.STRIDE_1X1, inplace_relu=cfg.RESNET.INPLACE_RELU, dilation=cfg.RESNET.SPATIAL_DILATIONS[0], norm_module=self.norm_module, ) # Based on profiling data of activation size, s1 and s2 have the activation sizes # that are 4X larger than the second largest. Therefore, checkpointing them gives # best memory savings. Further tuning is possible for better memory saving and tradeoffs # with recomputing FLOPs. if cfg.MODEL.ACT_CHECKPOINT: validate_checkpoint_wrapper_import(checkpoint_wrapper) self.s1 = checkpoint_wrapper(s1) self.s2 = checkpoint_wrapper(s2) else: self.s1 = s1 self.s2 = s2 for pathway in range(self.num_pathways): pool = nn.MaxPool3d( kernel_size=pool_size[pathway], stride=pool_size[pathway], padding=[0, 0, 0], ) self.add_module("pathway{}_pool".format(pathway), pool) self.s3 = resnet_helper.ResStage( dim_in=[width_per_group * 4], dim_out=[width_per_group * 8], dim_inner=[dim_inner * 2], temp_kernel_sizes=temp_kernel[2], stride=cfg.RESNET.SPATIAL_STRIDES[1], num_blocks=[d3], num_groups=[num_groups], num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[1], nonlocal_inds=cfg.NONLOCAL.LOCATION[1], nonlocal_group=cfg.NONLOCAL.GROUP[1], nonlocal_pool=cfg.NONLOCAL.POOL[1], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, stride_1x1=cfg.RESNET.STRIDE_1X1, inplace_relu=cfg.RESNET.INPLACE_RELU, dilation=cfg.RESNET.SPATIAL_DILATIONS[1], norm_module=self.norm_module, ) self.s4 = resnet_helper.ResStage( dim_in=[width_per_group * 8], dim_out=[width_per_group * 16], dim_inner=[dim_inner * 4], temp_kernel_sizes=temp_kernel[3], stride=cfg.RESNET.SPATIAL_STRIDES[2], num_blocks=[d4], num_groups=[num_groups], num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[2], nonlocal_inds=cfg.NONLOCAL.LOCATION[2], nonlocal_group=cfg.NONLOCAL.GROUP[2], nonlocal_pool=cfg.NONLOCAL.POOL[2], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, stride_1x1=cfg.RESNET.STRIDE_1X1, inplace_relu=cfg.RESNET.INPLACE_RELU, dilation=cfg.RESNET.SPATIAL_DILATIONS[2], norm_module=self.norm_module, ) self.s5 = resnet_helper.ResStage( dim_in=[width_per_group * 16], dim_out=[width_per_group * 32], dim_inner=[dim_inner * 8], temp_kernel_sizes=temp_kernel[4], stride=cfg.RESNET.SPATIAL_STRIDES[3], num_blocks=[d5], num_groups=[num_groups], num_block_temp_kernel=cfg.RESNET.NUM_BLOCK_TEMP_KERNEL[3], nonlocal_inds=cfg.NONLOCAL.LOCATION[3], nonlocal_group=cfg.NONLOCAL.GROUP[3], nonlocal_pool=cfg.NONLOCAL.POOL[3], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, stride_1x1=cfg.RESNET.STRIDE_1X1, inplace_relu=cfg.RESNET.INPLACE_RELU, dilation=cfg.RESNET.SPATIAL_DILATIONS[3], norm_module=self.norm_module, ) if self.enable_detection: self.head = head_helper.ResNetRoIHead( dim_in=[width_per_group * 32], num_classes=cfg.MODEL.NUM_CLASSES, pool_size=[[cfg.DATA.NUM_FRAMES // pool_size[0][0], 1, 1]], resolution=[[cfg.DETECTION.ROI_XFORM_RESOLUTION] * 2], scale_factor=[cfg.DETECTION.SPATIAL_SCALE_FACTOR], dropout_rate=cfg.MODEL.DROPOUT_RATE, act_func=cfg.MODEL.HEAD_ACT, aligned=cfg.DETECTION.ALIGNED, ) else: self.head = head_helper.ResNetBasicHead( dim_in=[width_per_group * 32], num_classes=cfg.MODEL.NUM_CLASSES, pool_size=[None, None] if cfg.MULTIGRID.SHORT_CYCLE else [ [ cfg.DATA.NUM_FRAMES // pool_size[0][0], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][1], cfg.DATA.TRAIN_CROP_SIZE // 32 // pool_size[0][2], ] ], # None for AdaptiveAvgPool3d((1, 1, 1)) dropout_rate=cfg.MODEL.DROPOUT_RATE, act_func=cfg.MODEL.HEAD_ACT, ) def forward(self, x, bboxes=None): x = self.s1(x) x = self.s2(x) y = [] # Don't modify x list in place due to activation checkpoint. for pathway in range(self.num_pathways): pool = getattr(self, "pathway{}_pool".format(pathway)) y.append(pool(x[pathway])) x = self.s3(y) x = self.s4(x) x = self.s5(x) if self.enable_detection: x = self.head(x, bboxes) else: x = self.head(x) return x @MODEL_REGISTRY.register() class X3D(nn.Module): """ X3D model builder. It builds a X3D network backbone, which is a ResNet. Christoph Feichtenhofer. "X3D: Expanding Architectures for Efficient Video Recognition." https://arxiv.org/abs/2004.04730 """ def __init__(self, cfg): """ The `__init__` method of any subclass should also contain these arguments. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ super(X3D, self).__init__() self.norm_module = get_norm(cfg) self.enable_detection = cfg.DETECTION.ENABLE self.num_pathways = 1 exp_stage = 2.0 self.dim_c1 = cfg.X3D.DIM_C1 self.dim_res2 = ( round_width(self.dim_c1, exp_stage, divisor=8) if cfg.X3D.SCALE_RES2 else self.dim_c1 ) self.dim_res3 = round_width(self.dim_res2, exp_stage, divisor=8) self.dim_res4 = round_width(self.dim_res3, exp_stage, divisor=8) self.dim_res5 = round_width(self.dim_res4, exp_stage, divisor=8) self.block_basis = [ # blocks, c, stride [1, self.dim_res2, 2], [2, self.dim_res3, 2], [5, self.dim_res4, 2], [3, self.dim_res5, 2], ] self._construct_network(cfg) init_helper.init_weights( self, cfg.MODEL.FC_INIT_STD, cfg.RESNET.ZERO_INIT_FINAL_BN ) def _round_repeats(self, repeats, multiplier): """Round number of layers based on depth multiplier.""" multiplier = multiplier if not multiplier: return repeats return int(math.ceil(multiplier * repeats)) def _construct_network(self, cfg): """ Builds a single pathway X3D model. Args: cfg (CfgNode): model building configs, details are in the comments of the config file. """ assert cfg.MODEL.ARCH in _POOL1.keys() assert cfg.RESNET.DEPTH in _MODEL_STAGE_DEPTH.keys() (d2, d3, d4, d5) = _MODEL_STAGE_DEPTH[cfg.RESNET.DEPTH] num_groups = cfg.RESNET.NUM_GROUPS width_per_group = cfg.RESNET.WIDTH_PER_GROUP dim_inner = num_groups * width_per_group w_mul = cfg.X3D.WIDTH_FACTOR d_mul = cfg.X3D.DEPTH_FACTOR dim_res1 = round_width(self.dim_c1, w_mul) temp_kernel = _TEMPORAL_KERNEL_BASIS[cfg.MODEL.ARCH] self.s1 = stem_helper.VideoModelStem( dim_in=cfg.DATA.INPUT_CHANNEL_NUM, dim_out=[dim_res1], kernel=[temp_kernel[0][0] + [3, 3]], stride=[[1, 2, 2]], padding=[[temp_kernel[0][0][0] // 2, 1, 1]], norm_module=self.norm_module, stem_func_name="x3d_stem", ) # blob_in = s1 dim_in = dim_res1 for stage, block in enumerate(self.block_basis): dim_out = round_width(block[1], w_mul) dim_inner = int(cfg.X3D.BOTTLENECK_FACTOR * dim_out) n_rep = self._round_repeats(block[0], d_mul) prefix = "s{}".format( stage + 2 ) # start w res2 to follow convention s = resnet_helper.ResStage( dim_in=[dim_in], dim_out=[dim_out], dim_inner=[dim_inner], temp_kernel_sizes=temp_kernel[1], stride=[block[2]], num_blocks=[n_rep], num_groups=[dim_inner] if cfg.X3D.CHANNELWISE_3x3x3 else [num_groups], num_block_temp_kernel=[n_rep], nonlocal_inds=cfg.NONLOCAL.LOCATION[0], nonlocal_group=cfg.NONLOCAL.GROUP[0], nonlocal_pool=cfg.NONLOCAL.POOL[0], instantiation=cfg.NONLOCAL.INSTANTIATION, trans_func_name=cfg.RESNET.TRANS_FUNC, stride_1x1=cfg.RESNET.STRIDE_1X1, norm_module=self.norm_module, dilation=cfg.RESNET.SPATIAL_DILATIONS[stage], drop_connect_rate=cfg.MODEL.DROPCONNECT_RATE * (stage + 2) / (len(self.block_basis) + 1), ) dim_in = dim_out self.add_module(prefix, s) if self.enable_detection: NotImplementedError else: spat_sz = int(math.ceil(cfg.DATA.TRAIN_CROP_SIZE / 32.0)) self.head = head_helper.X3DHead( dim_in=dim_out, dim_inner=dim_inner, dim_out=cfg.X3D.DIM_C5, num_classes=cfg.MODEL.NUM_CLASSES, pool_size=[cfg.DATA.NUM_FRAMES, spat_sz, spat_sz], dropout_rate=cfg.MODEL.DROPOUT_RATE, act_func=cfg.MODEL.HEAD_ACT, bn_lin5_on=cfg.X3D.BN_LIN5, ) def forward(self, x, bboxes=None): for module in self.children(): x = module(x) return x
28,931
36.044814
96
py
STTS
STTS-main/MViT/slowfast/models/common.py
# Copyright (c) Facebook, Inc. and its affiliates. import torch import torch.nn as nn class Mlp(nn.Module): def __init__( self, in_features, hidden_features=None, out_features=None, act_layer=nn.GELU, drop_rate=0.0, ): super().__init__() self.drop_rate = drop_rate out_features = out_features or in_features hidden_features = hidden_features or in_features self.fc1 = nn.Linear(in_features, hidden_features) self.act = act_layer() self.fc2 = nn.Linear(hidden_features, out_features) if self.drop_rate > 0.0: self.drop = nn.Dropout(drop_rate) def forward(self, x): x = self.fc1(x) x = self.act(x) if self.drop_rate > 0.0: x = self.drop(x) x = self.fc2(x) if self.drop_rate > 0.0: x = self.drop(x) return x class Permute(nn.Module): def __init__(self, dims): super().__init__() self.dims = dims def forward(self, x): return x.permute(*self.dims) def drop_path(x, drop_prob: float = 0.0, training: bool = False): """ Stochastic Depth per sample. """ if drop_prob == 0.0 or not training: return x keep_prob = 1 - drop_prob shape = (x.shape[0],) + (1,) * ( x.ndim - 1 ) # work with diff dim tensors, not just 2D ConvNets mask = keep_prob + torch.rand(shape, dtype=x.dtype, device=x.device) mask.floor_() # binarize output = x.div(keep_prob) * mask return output class DropPath(nn.Module): """Drop paths (Stochastic Depth) per sample (when applied in main path of residual blocks).""" def __init__(self, drop_prob=None): super(DropPath, self).__init__() self.drop_prob = drop_prob def forward(self, x): return drop_path(x, self.drop_prob, self.training)
1,900
25.774648
99
py
STTS
STTS-main/MViT/slowfast/models/head_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """ResNe(X)t Head helper.""" import torch import torch.nn as nn from detectron2.layers import ROIAlign class ResNetRoIHead(nn.Module): """ ResNe(X)t RoI head. """ def __init__( self, dim_in, num_classes, pool_size, resolution, scale_factor, dropout_rate=0.0, act_func="softmax", aligned=True, ): """ The `__init__` method of any subclass should also contain these arguments. ResNetRoIHead takes p pathways as input where p in [1, infty]. Args: dim_in (list): the list of channel dimensions of the p inputs to the ResNetHead. num_classes (int): the channel dimensions of the p outputs to the ResNetHead. pool_size (list): the list of kernel sizes of p spatial temporal poolings, temporal pool kernel size, spatial pool kernel size, spatial pool kernel size in order. resolution (list): the list of spatial output size from the ROIAlign. scale_factor (list): the list of ratio to the input boxes by this number. dropout_rate (float): dropout rate. If equal to 0.0, perform no dropout. act_func (string): activation function to use. 'softmax': applies softmax on the output. 'sigmoid': applies sigmoid on the output. aligned (bool): if False, use the legacy implementation. If True, align the results more perfectly. Note: Given a continuous coordinate c, its two neighboring pixel indices (in our pixel model) are computed by floor (c - 0.5) and ceil (c - 0.5). For example, c=1.3 has pixel neighbors with discrete indices [0] and [1] (which are sampled from the underlying signal at continuous coordinates 0.5 and 1.5). But the original roi_align (aligned=False) does not subtract the 0.5 when computing neighboring pixel indices and therefore it uses pixels with a slightly incorrect alignment (relative to our pixel model) when performing bilinear interpolation. With `aligned=True`, we first appropriately scale the ROI and then shift it by -0.5 prior to calling roi_align. This produces the correct neighbors; It makes negligible differences to the model's performance if ROIAlign is used together with conv layers. """ super(ResNetRoIHead, self).__init__() assert ( len({len(pool_size), len(dim_in)}) == 1 ), "pathway dimensions are not consistent." self.num_pathways = len(pool_size) for pathway in range(self.num_pathways): temporal_pool = nn.AvgPool3d( [pool_size[pathway][0], 1, 1], stride=1 ) self.add_module("s{}_tpool".format(pathway), temporal_pool) roi_align = ROIAlign( resolution[pathway], spatial_scale=1.0 / scale_factor[pathway], sampling_ratio=0, aligned=aligned, ) self.add_module("s{}_roi".format(pathway), roi_align) spatial_pool = nn.MaxPool2d(resolution[pathway], stride=1) self.add_module("s{}_spool".format(pathway), spatial_pool) if dropout_rate > 0.0: self.dropout = nn.Dropout(dropout_rate) # Perform FC in a fully convolutional manner. The FC layer will be # initialized with a different std comparing to convolutional layers. self.projection = nn.Linear(sum(dim_in), num_classes, bias=True) # Softmax for evaluation and testing. if act_func == "softmax": self.act = nn.Softmax(dim=1) elif act_func == "sigmoid": self.act = nn.Sigmoid() else: raise NotImplementedError( "{} is not supported as an activation" "function.".format(act_func) ) def forward(self, inputs, bboxes): assert ( len(inputs) == self.num_pathways ), "Input tensor does not contain {} pathway".format(self.num_pathways) pool_out = [] for pathway in range(self.num_pathways): t_pool = getattr(self, "s{}_tpool".format(pathway)) out = t_pool(inputs[pathway]) assert out.shape[2] == 1 out = torch.squeeze(out, 2) roi_align = getattr(self, "s{}_roi".format(pathway)) out = roi_align(out, bboxes) s_pool = getattr(self, "s{}_spool".format(pathway)) pool_out.append(s_pool(out)) # B C H W. x = torch.cat(pool_out, 1) # Perform dropout. if hasattr(self, "dropout"): x = self.dropout(x) x = x.view(x.shape[0], -1) x = self.projection(x) x = self.act(x) return x class ResNetBasicHead(nn.Module): """ ResNe(X)t 3D head. This layer performs a fully-connected projection during training, when the input size is 1x1x1. It performs a convolutional projection during testing when the input size is larger than 1x1x1. If the inputs are from multiple different pathways, the inputs will be concatenated after pooling. """ def __init__( self, dim_in, num_classes, pool_size, dropout_rate=0.0, act_func="softmax", ): """ The `__init__` method of any subclass should also contain these arguments. ResNetBasicHead takes p pathways as input where p in [1, infty]. Args: dim_in (list): the list of channel dimensions of the p inputs to the ResNetHead. num_classes (int): the channel dimensions of the p outputs to the ResNetHead. pool_size (list): the list of kernel sizes of p spatial temporal poolings, temporal pool kernel size, spatial pool kernel size, spatial pool kernel size in order. dropout_rate (float): dropout rate. If equal to 0.0, perform no dropout. act_func (string): activation function to use. 'softmax': applies softmax on the output. 'sigmoid': applies sigmoid on the output. """ super(ResNetBasicHead, self).__init__() assert ( len({len(pool_size), len(dim_in)}) == 1 ), "pathway dimensions are not consistent." self.num_pathways = len(pool_size) for pathway in range(self.num_pathways): if pool_size[pathway] is None: avg_pool = nn.AdaptiveAvgPool3d((1, 1, 1)) else: avg_pool = nn.AvgPool3d(pool_size[pathway], stride=1) self.add_module("pathway{}_avgpool".format(pathway), avg_pool) if dropout_rate > 0.0: self.dropout = nn.Dropout(dropout_rate) # Perform FC in a fully convolutional manner. The FC layer will be # initialized with a different std comparing to convolutional layers. self.projection = nn.Linear(sum(dim_in), num_classes, bias=True) # Softmax for evaluation and testing. if act_func == "softmax": self.act = nn.Softmax(dim=4) elif act_func == "sigmoid": self.act = nn.Sigmoid() else: raise NotImplementedError( "{} is not supported as an activation" "function.".format(act_func) ) def forward(self, inputs): assert ( len(inputs) == self.num_pathways ), "Input tensor does not contain {} pathway".format(self.num_pathways) pool_out = [] for pathway in range(self.num_pathways): m = getattr(self, "pathway{}_avgpool".format(pathway)) pool_out.append(m(inputs[pathway])) x = torch.cat(pool_out, 1) # (N, C, T, H, W) -> (N, T, H, W, C). x = x.permute((0, 2, 3, 4, 1)) # Perform dropout. if hasattr(self, "dropout"): x = self.dropout(x) x = self.projection(x) # Performs fully convlutional inference. if not self.training: x = self.act(x) x = x.mean([1, 2, 3]) x = x.view(x.shape[0], -1) return x class X3DHead(nn.Module): """ X3D head. This layer performs a fully-connected projection during training, when the input size is 1x1x1. It performs a convolutional projection during testing when the input size is larger than 1x1x1. If the inputs are from multiple different pathways, the inputs will be concatenated after pooling. """ def __init__( self, dim_in, dim_inner, dim_out, num_classes, pool_size, dropout_rate=0.0, act_func="softmax", inplace_relu=True, eps=1e-5, bn_mmt=0.1, norm_module=nn.BatchNorm3d, bn_lin5_on=False, ): """ The `__init__` method of any subclass should also contain these arguments. X3DHead takes a 5-dim feature tensor (BxCxTxHxW) as input. Args: dim_in (float): the channel dimension C of the input. num_classes (int): the channel dimensions of the output. pool_size (float): a single entry list of kernel size for spatiotemporal pooling for the TxHxW dimensions. dropout_rate (float): dropout rate. If equal to 0.0, perform no dropout. act_func (string): activation function to use. 'softmax': applies softmax on the output. 'sigmoid': applies sigmoid on the output. inplace_relu (bool): if True, calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. bn_lin5_on (bool): if True, perform normalization on the features before the classifier. """ super(X3DHead, self).__init__() self.pool_size = pool_size self.dropout_rate = dropout_rate self.num_classes = num_classes self.act_func = act_func self.eps = eps self.bn_mmt = bn_mmt self.inplace_relu = inplace_relu self.bn_lin5_on = bn_lin5_on self._construct_head(dim_in, dim_inner, dim_out, norm_module) def _construct_head(self, dim_in, dim_inner, dim_out, norm_module): self.conv_5 = nn.Conv3d( dim_in, dim_inner, kernel_size=(1, 1, 1), stride=(1, 1, 1), padding=(0, 0, 0), bias=False, ) self.conv_5_bn = norm_module( num_features=dim_inner, eps=self.eps, momentum=self.bn_mmt ) self.conv_5_relu = nn.ReLU(self.inplace_relu) if self.pool_size is None: self.avg_pool = nn.AdaptiveAvgPool3d((1, 1, 1)) else: self.avg_pool = nn.AvgPool3d(self.pool_size, stride=1) self.lin_5 = nn.Conv3d( dim_inner, dim_out, kernel_size=(1, 1, 1), stride=(1, 1, 1), padding=(0, 0, 0), bias=False, ) if self.bn_lin5_on: self.lin_5_bn = norm_module( num_features=dim_out, eps=self.eps, momentum=self.bn_mmt ) self.lin_5_relu = nn.ReLU(self.inplace_relu) if self.dropout_rate > 0.0: self.dropout = nn.Dropout(self.dropout_rate) # Perform FC in a fully convolutional manner. The FC layer will be # initialized with a different std comparing to convolutional layers. self.projection = nn.Linear(dim_out, self.num_classes, bias=True) # Softmax for evaluation and testing. if self.act_func == "softmax": self.act = nn.Softmax(dim=4) elif self.act_func == "sigmoid": self.act = nn.Sigmoid() else: raise NotImplementedError( "{} is not supported as an activation" "function.".format(self.act_func) ) def forward(self, inputs): # In its current design the X3D head is only useable for a single # pathway input. assert len(inputs) == 1, "Input tensor does not contain 1 pathway" x = self.conv_5(inputs[0]) x = self.conv_5_bn(x) x = self.conv_5_relu(x) x = self.avg_pool(x) x = self.lin_5(x) if self.bn_lin5_on: x = self.lin_5_bn(x) x = self.lin_5_relu(x) # (N, C, T, H, W) -> (N, T, H, W, C). x = x.permute((0, 2, 3, 4, 1)) # Perform dropout. if hasattr(self, "dropout"): x = self.dropout(x) x = self.projection(x) # Performs fully convlutional inference. if not self.training: x = self.act(x) x = x.mean([1, 2, 3]) x = x.view(x.shape[0], -1) return x class TransformerBasicHead(nn.Module): """ BasicHead. No pool. """ def __init__( self, dim_in, num_classes, dropout_rate=0.0, act_func="softmax", ): """ Perform linear projection and activation as head for tranformers. Args: dim_in (int): the channel dimension of the input to the head. num_classes (int): the channel dimensions of the output to the head. dropout_rate (float): dropout rate. If equal to 0.0, perform no dropout. act_func (string): activation function to use. 'softmax': applies softmax on the output. 'sigmoid': applies sigmoid on the output. """ super(TransformerBasicHead, self).__init__() if dropout_rate > 0.0: self.dropout = nn.Dropout(dropout_rate) self.projection = nn.Linear(dim_in, num_classes, bias=True) # Softmax for evaluation and testing. if act_func == "softmax": self.act = nn.Softmax(dim=1) elif act_func == "sigmoid": self.act = nn.Sigmoid() else: raise NotImplementedError( "{} is not supported as an activation" "function.".format(act_func) ) def forward(self, x): if hasattr(self, "dropout"): x = self.dropout(x) x = self.projection(x) if not self.training: x = self.act(x) return x
14,978
35.623472
81
py
STTS
STTS-main/MViT/slowfast/models/stem_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """ResNe(X)t 3D stem helper.""" import torch.nn as nn def get_stem_func(name): """ Retrieves the stem module by name. """ trans_funcs = {"x3d_stem": X3DStem, "basic_stem": ResNetBasicStem} assert ( name in trans_funcs.keys() ), "Transformation function '{}' not supported".format(name) return trans_funcs[name] class VideoModelStem(nn.Module): """ Video 3D stem module. Provides stem operations of Conv, BN, ReLU, MaxPool on input data tensor for one or multiple pathways. """ def __init__( self, dim_in, dim_out, kernel, stride, padding, inplace_relu=True, eps=1e-5, bn_mmt=0.1, norm_module=nn.BatchNorm3d, stem_func_name="basic_stem", ): """ The `__init__` method of any subclass should also contain these arguments. List size of 1 for single pathway models (C2D, I3D, Slow and etc), list size of 2 for two pathway models (SlowFast). Args: dim_in (list): the list of channel dimensions of the inputs. dim_out (list): the output dimension of the convolution in the stem layer. kernel (list): the kernels' size of the convolutions in the stem layers. Temporal kernel size, height kernel size, width kernel size in order. stride (list): the stride sizes of the convolutions in the stem layer. Temporal kernel stride, height kernel size, width kernel size in order. padding (list): the paddings' sizes of the convolutions in the stem layer. Temporal padding size, height padding size, width padding size in order. inplace_relu (bool): calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. stem_func_name (string): name of the the stem function applied on input to the network. """ super(VideoModelStem, self).__init__() assert ( len( { len(dim_in), len(dim_out), len(kernel), len(stride), len(padding), } ) == 1 ), "Input pathway dimensions are not consistent. {} {} {} {} {}".format( len(dim_in), len(dim_out), len(kernel), len(stride), len(padding), ) self.num_pathways = len(dim_in) self.kernel = kernel self.stride = stride self.padding = padding self.inplace_relu = inplace_relu self.eps = eps self.bn_mmt = bn_mmt # Construct the stem layer. self._construct_stem(dim_in, dim_out, norm_module, stem_func_name) def _construct_stem(self, dim_in, dim_out, norm_module, stem_func_name): trans_func = get_stem_func(stem_func_name) for pathway in range(len(dim_in)): stem = trans_func( dim_in[pathway], dim_out[pathway], self.kernel[pathway], self.stride[pathway], self.padding[pathway], self.inplace_relu, self.eps, self.bn_mmt, norm_module, ) self.add_module("pathway{}_stem".format(pathway), stem) def forward(self, x): assert ( len(x) == self.num_pathways ), "Input tensor does not contain {} pathway".format(self.num_pathways) # use a new list, don't modify in-place the x list, which is bad for activation checkpointing. y = [] for pathway in range(len(x)): m = getattr(self, "pathway{}_stem".format(pathway)) y.append(m(x[pathway])) return y class ResNetBasicStem(nn.Module): """ ResNe(X)t 3D stem module. Performs spatiotemporal Convolution, BN, and Relu following by a spatiotemporal pooling. """ def __init__( self, dim_in, dim_out, kernel, stride, padding, inplace_relu=True, eps=1e-5, bn_mmt=0.1, norm_module=nn.BatchNorm3d, ): """ The `__init__` method of any subclass should also contain these arguments. Args: dim_in (int): the channel dimension of the input. Normally 3 is used for rgb input, and 2 or 3 is used for optical flow input. dim_out (int): the output dimension of the convolution in the stem layer. kernel (list): the kernel size of the convolution in the stem layer. temporal kernel size, height kernel size, width kernel size in order. stride (list): the stride size of the convolution in the stem layer. temporal kernel stride, height kernel size, width kernel size in order. padding (int): the padding size of the convolution in the stem layer, temporal padding size, height padding size, width padding size in order. inplace_relu (bool): calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. """ super(ResNetBasicStem, self).__init__() self.kernel = kernel self.stride = stride self.padding = padding self.inplace_relu = inplace_relu self.eps = eps self.bn_mmt = bn_mmt # Construct the stem layer. self._construct_stem(dim_in, dim_out, norm_module) def _construct_stem(self, dim_in, dim_out, norm_module): self.conv = nn.Conv3d( dim_in, dim_out, self.kernel, stride=self.stride, padding=self.padding, bias=False, ) self.bn = norm_module( num_features=dim_out, eps=self.eps, momentum=self.bn_mmt ) self.relu = nn.ReLU(self.inplace_relu) self.pool_layer = nn.MaxPool3d( kernel_size=[1, 3, 3], stride=[1, 2, 2], padding=[0, 1, 1] ) def forward(self, x): x = self.conv(x) x = self.bn(x) x = self.relu(x) x = self.pool_layer(x) return x class X3DStem(nn.Module): """ X3D's 3D stem module. Performs a spatial followed by a depthwise temporal Convolution, BN, and Relu following by a spatiotemporal pooling. """ def __init__( self, dim_in, dim_out, kernel, stride, padding, inplace_relu=True, eps=1e-5, bn_mmt=0.1, norm_module=nn.BatchNorm3d, ): """ The `__init__` method of any subclass should also contain these arguments. Args: dim_in (int): the channel dimension of the input. Normally 3 is used for rgb input, and 2 or 3 is used for optical flow input. dim_out (int): the output dimension of the convolution in the stem layer. kernel (list): the kernel size of the convolution in the stem layer. temporal kernel size, height kernel size, width kernel size in order. stride (list): the stride size of the convolution in the stem layer. temporal kernel stride, height kernel size, width kernel size in order. padding (int): the padding size of the convolution in the stem layer, temporal padding size, height padding size, width padding size in order. inplace_relu (bool): calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. """ super(X3DStem, self).__init__() self.kernel = kernel self.stride = stride self.padding = padding self.inplace_relu = inplace_relu self.eps = eps self.bn_mmt = bn_mmt # Construct the stem layer. self._construct_stem(dim_in, dim_out, norm_module) def _construct_stem(self, dim_in, dim_out, norm_module): self.conv_xy = nn.Conv3d( dim_in, dim_out, kernel_size=(1, self.kernel[1], self.kernel[2]), stride=(1, self.stride[1], self.stride[2]), padding=(0, self.padding[1], self.padding[2]), bias=False, ) self.conv = nn.Conv3d( dim_out, dim_out, kernel_size=(self.kernel[0], 1, 1), stride=(self.stride[0], 1, 1), padding=(self.padding[0], 0, 0), bias=False, groups=dim_out, ) self.bn = norm_module( num_features=dim_out, eps=self.eps, momentum=self.bn_mmt ) self.relu = nn.ReLU(self.inplace_relu) def forward(self, x): x = self.conv_xy(x) x = self.conv(x) x = self.bn(x) x = self.relu(x) return x class PatchEmbed(nn.Module): """ PatchEmbed. """ def __init__( self, dim_in=3, dim_out=768, kernel=(1, 16, 16), stride=(1, 4, 4), padding=(1, 7, 7), conv_2d=False, ): super().__init__() if conv_2d: conv = nn.Conv2d else: conv = nn.Conv3d self.proj = conv( dim_in, dim_out, kernel_size=kernel, stride=stride, padding=padding, ) def forward(self, x): x = self.proj(x) # B C (T) H W -> B (T)HW C return x.flatten(2).transpose(1, 2)
10,775
32.362229
102
py
STTS
STTS-main/MViT/slowfast/models/__init__.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. from .build import MODEL_REGISTRY, build_model # noqa from .custom_video_model_builder import * # noqa from .video_model_builder import ResNet, SlowFast # noqa from .mvit import MViT try: from .ptv_model_builder import ( PTVCSN, PTVX3D, PTVR2plus1D, PTVResNet, PTVSlowFast, ) # noqa except Exception: print("Please update your PyTorchVideo to latest master")
513
24.7
71
py
STTS
STTS-main/MViT/slowfast/models/resnet_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Video models.""" import torch import torch.nn as nn from slowfast.models.common import drop_path from slowfast.models.nonlocal_helper import Nonlocal from slowfast.models.operators import SE, Swish def get_trans_func(name): """ Retrieves the transformation module by name. """ trans_funcs = { "bottleneck_transform": BottleneckTransform, "basic_transform": BasicTransform, "x3d_transform": X3DTransform, } assert ( name in trans_funcs.keys() ), "Transformation function '{}' not supported".format(name) return trans_funcs[name] class BasicTransform(nn.Module): """ Basic transformation: Tx3x3, 1x3x3, where T is the size of temporal kernel. """ def __init__( self, dim_in, dim_out, temp_kernel_size, stride, dim_inner=None, num_groups=1, stride_1x1=None, inplace_relu=True, eps=1e-5, bn_mmt=0.1, norm_module=nn.BatchNorm3d, block_idx=0, ): """ Args: dim_in (int): the channel dimensions of the input. dim_out (int): the channel dimension of the output. temp_kernel_size (int): the temporal kernel sizes of the first convolution in the basic block. stride (int): the stride of the bottleneck. dim_inner (None): the inner dimension would not be used in BasicTransform. num_groups (int): number of groups for the convolution. Number of group is always 1 for BasicTransform. stride_1x1 (None): stride_1x1 will not be used in BasicTransform. inplace_relu (bool): if True, calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. """ super(BasicTransform, self).__init__() self.temp_kernel_size = temp_kernel_size self._inplace_relu = inplace_relu self._eps = eps self._bn_mmt = bn_mmt self._construct(dim_in, dim_out, stride, norm_module) def _construct(self, dim_in, dim_out, stride, norm_module): # Tx3x3, BN, ReLU. self.a = nn.Conv3d( dim_in, dim_out, kernel_size=[self.temp_kernel_size, 3, 3], stride=[1, stride, stride], padding=[int(self.temp_kernel_size // 2), 1, 1], bias=False, ) self.a_bn = norm_module( num_features=dim_out, eps=self._eps, momentum=self._bn_mmt ) self.a_relu = nn.ReLU(inplace=self._inplace_relu) # 1x3x3, BN. self.b = nn.Conv3d( dim_out, dim_out, kernel_size=[1, 3, 3], stride=[1, 1, 1], padding=[0, 1, 1], bias=False, ) self.b_bn = norm_module( num_features=dim_out, eps=self._eps, momentum=self._bn_mmt ) self.b_bn.transform_final_bn = True def forward(self, x): x = self.a(x) x = self.a_bn(x) x = self.a_relu(x) x = self.b(x) x = self.b_bn(x) return x class X3DTransform(nn.Module): """ X3D transformation: 1x1x1, Tx3x3 (channelwise, num_groups=dim_in), 1x1x1, augmented with (optional) SE (squeeze-excitation) on the 3x3x3 output. T is the temporal kernel size (defaulting to 3) """ def __init__( self, dim_in, dim_out, temp_kernel_size, stride, dim_inner, num_groups, stride_1x1=False, inplace_relu=True, eps=1e-5, bn_mmt=0.1, dilation=1, norm_module=nn.BatchNorm3d, se_ratio=0.0625, swish_inner=True, block_idx=0, ): """ Args: dim_in (int): the channel dimensions of the input. dim_out (int): the channel dimension of the output. temp_kernel_size (int): the temporal kernel sizes of the middle convolution in the bottleneck. stride (int): the stride of the bottleneck. dim_inner (int): the inner dimension of the block. num_groups (int): number of groups for the convolution. num_groups=1 is for standard ResNet like networks, and num_groups>1 is for ResNeXt like networks. stride_1x1 (bool): if True, apply stride to 1x1 conv, otherwise apply stride to the 3x3 conv. inplace_relu (bool): if True, calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. dilation (int): size of dilation. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. se_ratio (float): if > 0, apply SE to the Tx3x3 conv, with the SE channel dimensionality being se_ratio times the Tx3x3 conv dim. swish_inner (bool): if True, apply swish to the Tx3x3 conv, otherwise apply ReLU to the Tx3x3 conv. """ super(X3DTransform, self).__init__() self.temp_kernel_size = temp_kernel_size self._inplace_relu = inplace_relu self._eps = eps self._bn_mmt = bn_mmt self._se_ratio = se_ratio self._swish_inner = swish_inner self._stride_1x1 = stride_1x1 self._block_idx = block_idx self._construct( dim_in, dim_out, stride, dim_inner, num_groups, dilation, norm_module, ) def _construct( self, dim_in, dim_out, stride, dim_inner, num_groups, dilation, norm_module, ): (str1x1, str3x3) = (stride, 1) if self._stride_1x1 else (1, stride) # 1x1x1, BN, ReLU. self.a = nn.Conv3d( dim_in, dim_inner, kernel_size=[1, 1, 1], stride=[1, str1x1, str1x1], padding=[0, 0, 0], bias=False, ) self.a_bn = norm_module( num_features=dim_inner, eps=self._eps, momentum=self._bn_mmt ) self.a_relu = nn.ReLU(inplace=self._inplace_relu) # Tx3x3, BN, ReLU. self.b = nn.Conv3d( dim_inner, dim_inner, [self.temp_kernel_size, 3, 3], stride=[1, str3x3, str3x3], padding=[int(self.temp_kernel_size // 2), dilation, dilation], groups=num_groups, bias=False, dilation=[1, dilation, dilation], ) self.b_bn = norm_module( num_features=dim_inner, eps=self._eps, momentum=self._bn_mmt ) # Apply SE attention or not use_se = True if (self._block_idx + 1) % 2 else False if self._se_ratio > 0.0 and use_se: self.se = SE(dim_inner, self._se_ratio) if self._swish_inner: self.b_relu = Swish() else: self.b_relu = nn.ReLU(inplace=self._inplace_relu) # 1x1x1, BN. self.c = nn.Conv3d( dim_inner, dim_out, kernel_size=[1, 1, 1], stride=[1, 1, 1], padding=[0, 0, 0], bias=False, ) self.c_bn = norm_module( num_features=dim_out, eps=self._eps, momentum=self._bn_mmt ) self.c_bn.transform_final_bn = True def forward(self, x): for block in self.children(): x = block(x) return x class BottleneckTransform(nn.Module): """ Bottleneck transformation: Tx1x1, 1x3x3, 1x1x1, where T is the size of temporal kernel. """ def __init__( self, dim_in, dim_out, temp_kernel_size, stride, dim_inner, num_groups, stride_1x1=False, inplace_relu=True, eps=1e-5, bn_mmt=0.1, dilation=1, norm_module=nn.BatchNorm3d, block_idx=0, ): """ Args: dim_in (int): the channel dimensions of the input. dim_out (int): the channel dimension of the output. temp_kernel_size (int): the temporal kernel sizes of the first convolution in the bottleneck. stride (int): the stride of the bottleneck. dim_inner (int): the inner dimension of the block. num_groups (int): number of groups for the convolution. num_groups=1 is for standard ResNet like networks, and num_groups>1 is for ResNeXt like networks. stride_1x1 (bool): if True, apply stride to 1x1 conv, otherwise apply stride to the 3x3 conv. inplace_relu (bool): if True, calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. dilation (int): size of dilation. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. """ super(BottleneckTransform, self).__init__() self.temp_kernel_size = temp_kernel_size self._inplace_relu = inplace_relu self._eps = eps self._bn_mmt = bn_mmt self._stride_1x1 = stride_1x1 self._construct( dim_in, dim_out, stride, dim_inner, num_groups, dilation, norm_module, ) def _construct( self, dim_in, dim_out, stride, dim_inner, num_groups, dilation, norm_module, ): (str1x1, str3x3) = (stride, 1) if self._stride_1x1 else (1, stride) # Tx1x1, BN, ReLU. self.a = nn.Conv3d( dim_in, dim_inner, kernel_size=[self.temp_kernel_size, 1, 1], stride=[1, str1x1, str1x1], padding=[int(self.temp_kernel_size // 2), 0, 0], bias=False, ) self.a_bn = norm_module( num_features=dim_inner, eps=self._eps, momentum=self._bn_mmt ) self.a_relu = nn.ReLU(inplace=self._inplace_relu) # 1x3x3, BN, ReLU. self.b = nn.Conv3d( dim_inner, dim_inner, [1, 3, 3], stride=[1, str3x3, str3x3], padding=[0, dilation, dilation], groups=num_groups, bias=False, dilation=[1, dilation, dilation], ) self.b_bn = norm_module( num_features=dim_inner, eps=self._eps, momentum=self._bn_mmt ) self.b_relu = nn.ReLU(inplace=self._inplace_relu) # 1x1x1, BN. self.c = nn.Conv3d( dim_inner, dim_out, kernel_size=[1, 1, 1], stride=[1, 1, 1], padding=[0, 0, 0], bias=False, ) self.c_bn = norm_module( num_features=dim_out, eps=self._eps, momentum=self._bn_mmt ) self.c_bn.transform_final_bn = True def forward(self, x): # Explicitly forward every layer. # Branch2a. x = self.a(x) x = self.a_bn(x) x = self.a_relu(x) # Branch2b. x = self.b(x) x = self.b_bn(x) x = self.b_relu(x) # Branch2c x = self.c(x) x = self.c_bn(x) return x class ResBlock(nn.Module): """ Residual block. """ def __init__( self, dim_in, dim_out, temp_kernel_size, stride, trans_func, dim_inner, num_groups=1, stride_1x1=False, inplace_relu=True, eps=1e-5, bn_mmt=0.1, dilation=1, norm_module=nn.BatchNorm3d, block_idx=0, drop_connect_rate=0.0, ): """ ResBlock class constructs redisual blocks. More details can be found in: Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. "Deep residual learning for image recognition." https://arxiv.org/abs/1512.03385 Args: dim_in (int): the channel dimensions of the input. dim_out (int): the channel dimension of the output. temp_kernel_size (int): the temporal kernel sizes of the middle convolution in the bottleneck. stride (int): the stride of the bottleneck. trans_func (string): transform function to be used to construct the bottleneck. dim_inner (int): the inner dimension of the block. num_groups (int): number of groups for the convolution. num_groups=1 is for standard ResNet like networks, and num_groups>1 is for ResNeXt like networks. stride_1x1 (bool): if True, apply stride to 1x1 conv, otherwise apply stride to the 3x3 conv. inplace_relu (bool): calculate the relu on the original input without allocating new memory. eps (float): epsilon for batch norm. bn_mmt (float): momentum for batch norm. Noted that BN momentum in PyTorch = 1 - BN momentum in Caffe2. dilation (int): size of dilation. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. drop_connect_rate (float): basic rate at which blocks are dropped, linearly increases from input to output blocks. """ super(ResBlock, self).__init__() self._inplace_relu = inplace_relu self._eps = eps self._bn_mmt = bn_mmt self._drop_connect_rate = drop_connect_rate self._construct( dim_in, dim_out, temp_kernel_size, stride, trans_func, dim_inner, num_groups, stride_1x1, inplace_relu, dilation, norm_module, block_idx, ) def _construct( self, dim_in, dim_out, temp_kernel_size, stride, trans_func, dim_inner, num_groups, stride_1x1, inplace_relu, dilation, norm_module, block_idx, ): # Use skip connection with projection if dim or res change. if (dim_in != dim_out) or (stride != 1): self.branch1 = nn.Conv3d( dim_in, dim_out, kernel_size=1, stride=[1, stride, stride], padding=0, bias=False, dilation=1, ) self.branch1_bn = norm_module( num_features=dim_out, eps=self._eps, momentum=self._bn_mmt ) self.branch2 = trans_func( dim_in, dim_out, temp_kernel_size, stride, dim_inner, num_groups, stride_1x1=stride_1x1, inplace_relu=inplace_relu, dilation=dilation, norm_module=norm_module, block_idx=block_idx, ) self.relu = nn.ReLU(self._inplace_relu) def forward(self, x): f_x = self.branch2(x) if self.training and self._drop_connect_rate > 0.0: f_x = drop_path(f_x, self._drop_connect_rate) if hasattr(self, "branch1"): x = self.branch1_bn(self.branch1(x)) + f_x else: x = x + f_x x = self.relu(x) return x class ResStage(nn.Module): """ Stage of 3D ResNet. It expects to have one or more tensors as input for single pathway (C2D, I3D, Slow), and multi-pathway (SlowFast) cases. More details can be found here: Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. "SlowFast networks for video recognition." https://arxiv.org/pdf/1812.03982.pdf """ def __init__( self, dim_in, dim_out, stride, temp_kernel_sizes, num_blocks, dim_inner, num_groups, num_block_temp_kernel, nonlocal_inds, nonlocal_group, nonlocal_pool, dilation, instantiation="softmax", trans_func_name="bottleneck_transform", stride_1x1=False, inplace_relu=True, norm_module=nn.BatchNorm3d, drop_connect_rate=0.0, ): """ The `__init__` method of any subclass should also contain these arguments. ResStage builds p streams, where p can be greater or equal to one. Args: dim_in (list): list of p the channel dimensions of the input. Different channel dimensions control the input dimension of different pathways. dim_out (list): list of p the channel dimensions of the output. Different channel dimensions control the input dimension of different pathways. temp_kernel_sizes (list): list of the p temporal kernel sizes of the convolution in the bottleneck. Different temp_kernel_sizes control different pathway. stride (list): list of the p strides of the bottleneck. Different stride control different pathway. num_blocks (list): list of p numbers of blocks for each of the pathway. dim_inner (list): list of the p inner channel dimensions of the input. Different channel dimensions control the input dimension of different pathways. num_groups (list): list of number of p groups for the convolution. num_groups=1 is for standard ResNet like networks, and num_groups>1 is for ResNeXt like networks. num_block_temp_kernel (list): extent the temp_kernel_sizes to num_block_temp_kernel blocks, then fill temporal kernel size of 1 for the rest of the layers. nonlocal_inds (list): If the tuple is empty, no nonlocal layer will be added. If the tuple is not empty, add nonlocal layers after the index-th block. dilation (list): size of dilation for each pathway. nonlocal_group (list): list of number of p nonlocal groups. Each number controls how to fold temporal dimension to batch dimension before applying nonlocal transformation. https://github.com/facebookresearch/video-nonlocal-net. instantiation (string): different instantiation for nonlocal layer. Supports two different instantiation method: "dot_product": normalizing correlation matrix with L2. "softmax": normalizing correlation matrix with Softmax. trans_func_name (string): name of the the transformation function apply on the network. norm_module (nn.Module): nn.Module for the normalization layer. The default is nn.BatchNorm3d. drop_connect_rate (float): basic rate at which blocks are dropped, linearly increases from input to output blocks. """ super(ResStage, self).__init__() assert all( ( num_block_temp_kernel[i] <= num_blocks[i] for i in range(len(temp_kernel_sizes)) ) ) self.num_blocks = num_blocks self.nonlocal_group = nonlocal_group self._drop_connect_rate = drop_connect_rate self.temp_kernel_sizes = [ (temp_kernel_sizes[i] * num_blocks[i])[: num_block_temp_kernel[i]] + [1] * (num_blocks[i] - num_block_temp_kernel[i]) for i in range(len(temp_kernel_sizes)) ] assert ( len( { len(dim_in), len(dim_out), len(temp_kernel_sizes), len(stride), len(num_blocks), len(dim_inner), len(num_groups), len(num_block_temp_kernel), len(nonlocal_inds), len(nonlocal_group), } ) == 1 ) self.num_pathways = len(self.num_blocks) self._construct( dim_in, dim_out, stride, dim_inner, num_groups, trans_func_name, stride_1x1, inplace_relu, nonlocal_inds, nonlocal_pool, instantiation, dilation, norm_module, ) def _construct( self, dim_in, dim_out, stride, dim_inner, num_groups, trans_func_name, stride_1x1, inplace_relu, nonlocal_inds, nonlocal_pool, instantiation, dilation, norm_module, ): for pathway in range(self.num_pathways): for i in range(self.num_blocks[pathway]): # Retrieve the transformation function. trans_func = get_trans_func(trans_func_name) # Construct the block. res_block = ResBlock( dim_in[pathway] if i == 0 else dim_out[pathway], dim_out[pathway], self.temp_kernel_sizes[pathway][i], stride[pathway] if i == 0 else 1, trans_func, dim_inner[pathway], num_groups[pathway], stride_1x1=stride_1x1, inplace_relu=inplace_relu, dilation=dilation[pathway], norm_module=norm_module, block_idx=i, drop_connect_rate=self._drop_connect_rate, ) self.add_module("pathway{}_res{}".format(pathway, i), res_block) if i in nonlocal_inds[pathway]: nln = Nonlocal( dim_out[pathway], dim_out[pathway] // 2, nonlocal_pool[pathway], instantiation=instantiation, norm_module=norm_module, ) self.add_module( "pathway{}_nonlocal{}".format(pathway, i), nln ) def forward(self, inputs): output = [] for pathway in range(self.num_pathways): x = inputs[pathway] for i in range(self.num_blocks[pathway]): m = getattr(self, "pathway{}_res{}".format(pathway, i)) x = m(x) if hasattr(self, "pathway{}_nonlocal{}".format(pathway, i)): nln = getattr( self, "pathway{}_nonlocal{}".format(pathway, i) ) b, c, t, h, w = x.shape if self.nonlocal_group[pathway] > 1: # Fold temporal dimension into batch dimension. x = x.permute(0, 2, 1, 3, 4) x = x.reshape( b * self.nonlocal_group[pathway], t // self.nonlocal_group[pathway], c, h, w, ) x = x.permute(0, 2, 1, 3, 4) x = nln(x) if self.nonlocal_group[pathway] > 1: # Fold back to temporal dimension. x = x.permute(0, 2, 1, 3, 4) x = x.reshape(b, t, c, h, w) x = x.permute(0, 2, 1, 3, 4) output.append(x) return output
24,795
33.15427
83
py
STTS
STTS-main/MViT/slowfast/models/attention.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import numpy import torch import torch.nn as nn from einops import rearrange from slowfast.models.common import DropPath, Mlp def attention_pool(tensor, pool, thw_shape, has_cls_embed=True, norm=None): if pool is None: return tensor, thw_shape tensor_dim = tensor.ndim if tensor_dim == 4: pass elif tensor_dim == 3: tensor = tensor.unsqueeze(1) else: raise NotImplementedError(f"Unsupported input dimension {tensor.shape}") if has_cls_embed: cls_tok, tensor = tensor[:, :, :1, :], tensor[:, :, 1:, :] B, N, L, C = tensor.shape T, H, W = thw_shape tensor = ( tensor.reshape(B * N, T, H, W, C).permute(0, 4, 1, 2, 3).contiguous() ) tensor = pool(tensor) thw_shape = [tensor.shape[2], tensor.shape[3], tensor.shape[4]] L_pooled = tensor.shape[2] * tensor.shape[3] * tensor.shape[4] tensor = tensor.reshape(B, N, C, L_pooled).transpose(2, 3) if has_cls_embed: tensor = torch.cat((cls_tok, tensor), dim=2) if norm is not None: tensor = norm(tensor) # Assert tensor_dim in [3, 4] if tensor_dim == 4: pass else: # tensor_dim == 3: tensor = tensor.squeeze(1) return tensor, thw_shape class MultiScaleAttention(nn.Module): def __init__( self, dim, num_heads=8, qkv_bias=False, drop_rate=0.0, kernel_q=(1, 1, 1), kernel_kv=(1, 1, 1), stride_q=(1, 1, 1), stride_kv=(1, 1, 1), norm_layer=nn.LayerNorm, has_cls_embed=True, # Options include `conv`, `avg`, and `max`. mode="conv", # If True, perform pool before projection. pool_first=False, ): super().__init__() self.pool_first = pool_first self.drop_rate = drop_rate self.num_heads = num_heads head_dim = dim // num_heads self.scale = head_dim ** -0.5 self.has_cls_embed = has_cls_embed padding_q = [int(q // 2) for q in kernel_q] padding_kv = [int(kv // 2) for kv in kernel_kv] #self.q = nn.Linear(dim, dim, bias=qkv_bias) #self.k = nn.Linear(dim, dim, bias=qkv_bias) #self.v = nn.Linear(dim, dim, bias=qkv_bias) self.qkv = nn.Linear(dim, dim*3, bias=qkv_bias) self.proj = nn.Linear(dim, dim) if drop_rate > 0.0: self.proj_drop = nn.Dropout(drop_rate) # Skip pooling with kernel and stride size of (1, 1, 1). if numpy.prod(kernel_q) == 1 and numpy.prod(stride_q) == 1: kernel_q = () if numpy.prod(kernel_kv) == 1 and numpy.prod(stride_kv) == 1: kernel_kv = () if mode in ("avg", "max"): pool_op = nn.MaxPool3d if mode == "max" else nn.AvgPool3d self.pool_q = ( pool_op(kernel_q, stride_q, padding_q, ceil_mode=False) if len(kernel_q) > 0 else None ) self.pool_k = ( pool_op(kernel_kv, stride_kv, padding_kv, ceil_mode=False) if len(kernel_kv) > 0 else None ) self.pool_v = ( pool_op(kernel_kv, stride_kv, padding_kv, ceil_mode=False) if len(kernel_kv) > 0 else None ) elif mode == "conv": self.pool_q = ( nn.Conv3d( head_dim, head_dim, kernel_q, stride=stride_q, padding=padding_q, groups=head_dim, bias=False, ) if len(kernel_q) > 0 else None ) self.norm_q = norm_layer(head_dim) if len(kernel_q) > 0 else None self.pool_k = ( nn.Conv3d( head_dim, head_dim, kernel_kv, stride=stride_kv, padding=padding_kv, groups=head_dim, bias=False, ) if len(kernel_kv) > 0 else None ) self.norm_k = norm_layer(head_dim) if len(kernel_kv) > 0 else None self.pool_v = ( nn.Conv3d( head_dim, head_dim, kernel_kv, stride=stride_kv, padding=padding_kv, groups=head_dim, bias=False, ) if len(kernel_kv) > 0 else None ) # print(kernel_kv) self.norm_v = norm_layer(head_dim) if len(kernel_kv) > 0 else None else: raise NotImplementedError(f"Unsupported model {mode}") def forward(self, x, thw_shape): B, N, C = x.shape q = k = v = x qkv = ( self.qkv(x) .reshape(B, N, 3, self.num_heads, C // self.num_heads) .permute(2, 0, 3, 1, 4).contiguous() ) q, k, v = qkv[0], qkv[1], qkv[2] q, q_shape = attention_pool( q, self.pool_q, thw_shape, has_cls_embed=self.has_cls_embed, norm=self.norm_q if hasattr(self, "norm_q") else None, ) k, _ = attention_pool( k, self.pool_k, thw_shape, has_cls_embed=self.has_cls_embed, norm=self.norm_k if hasattr(self, "norm_k") else None, ) v, _ = attention_pool( v, self.pool_v, thw_shape, has_cls_embed=self.has_cls_embed, norm=self.norm_v if hasattr(self, "norm_v") else None, ) attn = (q @ k.transpose(-2, -1)) * self.scale attn = attn.softmax(dim=-1) N = q.shape[2] x = (attn @ v).transpose(1, 2).reshape(B, N, C) x = self.proj(x) if self.drop_rate > 0.0: x = self.proj_drop(x) return x, q_shape class MultiScaleBlock(nn.Module): def __init__( self, dim, dim_out, num_heads, mlp_ratio=4.0, qkv_bias=False, qk_scale=None, drop_rate=0.0, drop_path=0.0, act_layer=nn.GELU, norm_layer=nn.LayerNorm, up_rate=None, kernel_q=(1, 1, 1), kernel_kv=(1, 1, 1), stride_q=(1, 1, 1), stride_kv=(1, 1, 1), mode="conv", has_cls_embed=True, pool_first=False, ): super().__init__() self.dim = dim self.dim_out = dim_out self.norm1 = norm_layer(dim) kernel_skip = [s + 1 if s > 1 else s for s in stride_q] stride_skip = stride_q padding_skip = [int(skip // 2) for skip in kernel_skip] self.attn = MultiScaleAttention( dim, num_heads=num_heads, qkv_bias=qkv_bias, drop_rate=drop_rate, kernel_q=kernel_q, kernel_kv=kernel_kv, stride_q=stride_q, stride_kv=stride_kv, norm_layer=nn.LayerNorm, has_cls_embed=has_cls_embed, mode=mode, pool_first=pool_first, ) self.drop_path = ( DropPath(drop_path) if drop_path > 0.0 else nn.Identity() ) self.norm2 = norm_layer(dim) mlp_hidden_dim = int(dim * mlp_ratio) self.has_cls_embed = has_cls_embed # TODO: check the use case for up_rate, and merge the following lines if up_rate is not None and up_rate > 1: mlp_dim_out = dim * up_rate else: mlp_dim_out = dim_out self.mlp = Mlp( in_features=dim, hidden_features=mlp_hidden_dim, out_features=mlp_dim_out, act_layer=act_layer, drop_rate=drop_rate, ) if dim != dim_out: self.proj = nn.Linear(dim, dim_out) self.pool_skip = ( nn.MaxPool3d( kernel_skip, stride_skip, padding_skip, ceil_mode=False ) if len(kernel_skip) > 0 else None ) def forward(self, x, thw_shape): x_block, thw_shape_new = self.attn(self.norm1(x), thw_shape) x_res, _ = attention_pool( x, self.pool_skip, thw_shape, has_cls_embed=self.has_cls_embed ) x = x_res + self.drop_path(x_block) x_norm = self.norm2(x) x_mlp = self.mlp(x_norm) if self.dim != self.dim_out: x = self.proj(x_norm) x = x + self.drop_path(x_mlp) return x, thw_shape_new
8,875
29.712803
80
py
STTS
STTS-main/MViT/slowfast/models/build.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Model construction functions.""" import torch from fvcore.common.registry import Registry MODEL_REGISTRY = Registry("MODEL") MODEL_REGISTRY.__doc__ = """ Registry for video model. The registered object will be called with `obj(cfg)`. The call should return a `torch.nn.Module` object. """ def build_model(cfg, gpu_id=None): """ Builds the video model. Args: cfg (configs): configs that contains the hyper-parameters to build the backbone. Details can be seen in slowfast/config/defaults.py. gpu_id (Optional[int]): specify the gpu index to build model. """ if torch.cuda.is_available(): assert ( cfg.NUM_GPUS <= torch.cuda.device_count() ), "Cannot use more GPU devices than available" else: assert ( cfg.NUM_GPUS == 0 ), "Cuda is not available. Please set `NUM_GPUS: 0 for running on CPUs." # Construct the model name = cfg.MODEL.MODEL_NAME model = MODEL_REGISTRY.get(name)(cfg) if cfg.NUM_GPUS: if gpu_id is None: # Determine the GPU used by the current process cur_device = torch.cuda.current_device() else: cur_device = gpu_id # Transfer the model to the current GPU device model = model.cuda(device=cur_device) # Use multi-process data parallel model in the multi-gpu setting if cfg.NUM_GPUS > 1: # Make model replica operate on the current device model = torch.nn.parallel.DistributedDataParallel( module=model, device_ids=[cur_device], output_device=cur_device ) return model
1,725
30.962963
80
py
STTS
STTS-main/MViT/slowfast/models/optimizer.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Optimizer.""" import torch import slowfast.utils.lr_policy as lr_policy def construct_optimizer(model, cfg): """ Construct a stochastic gradient descent or ADAM optimizer with momentum. Details can be found in: Herbert Robbins, and Sutton Monro. "A stochastic approximation method." and Diederik P.Kingma, and Jimmy Ba. "Adam: A Method for Stochastic Optimization." Args: model (model): model to perform stochastic gradient descent optimization or ADAM optimization. cfg (config): configs of hyper-parameters of SGD or ADAM, includes base learning rate, momentum, weight_decay, dampening, and etc. """ train_topk_only = cfg.TRAIN.TRAIN_TOPK_ONLY # Batchnorm parameters. bn_params = [] # Non-batchnorm parameters. non_bn_params = [] zero_params = [] predictor = [] skip = {} if hasattr(model, "no_weight_decay"): skip = model.no_weight_decay() for name, m in model.named_modules(): is_bn = isinstance(m, torch.nn.modules.batchnorm._NormBase) for p in m.parameters(recurse=False): if not p.requires_grad: continue if 'predictor' in name: predictor.append(p) elif train_topk_only: continue elif is_bn: bn_params.append(p) elif name in skip or ( (len(p.shape) == 1 or name.endswith(".bias")) and cfg.SOLVER.ZERO_WD_1D_PARAM ): zero_params.append(p) else: non_bn_params.append(p) optim_params = [ {"params": predictor, "weight_decay": cfg.SOLVER.WEIGHT_DECAY, 'name': 'predictor'}, {"params": bn_params, "weight_decay": cfg.BN.WEIGHT_DECAY, 'name': 'backbone_bn'}, {"params": non_bn_params, "weight_decay": cfg.SOLVER.WEIGHT_DECAY, 'name': 'backbone_nonbn'}, {"params": zero_params, "weight_decay": 0.0, 'name': 'bacbone_zero'}, ] optim_params = [x for x in optim_params if len(x["params"])] if cfg.SOLVER.OPTIMIZING_METHOD == "sgd": return torch.optim.SGD( optim_params, lr=cfg.SOLVER.BASE_LR, momentum=cfg.SOLVER.MOMENTUM, weight_decay=cfg.SOLVER.WEIGHT_DECAY, dampening=cfg.SOLVER.DAMPENING, nesterov=cfg.SOLVER.NESTEROV, ) elif cfg.SOLVER.OPTIMIZING_METHOD == "adam": return torch.optim.Adam( optim_params, lr=cfg.SOLVER.BASE_LR, betas=(0.9, 0.999), eps=1e-08, weight_decay=cfg.SOLVER.WEIGHT_DECAY, ) elif cfg.SOLVER.OPTIMIZING_METHOD == "adamw": return torch.optim.AdamW( optim_params, lr=cfg.SOLVER.BASE_LR, betas=(0.9, 0.999), eps=1e-08, weight_decay=cfg.SOLVER.WEIGHT_DECAY, ) else: raise NotImplementedError( "Does not support {} optimizer".format(cfg.SOLVER.OPTIMIZING_METHOD) ) def get_epoch_lr(cur_epoch, cfg): """ Retrieves the lr for the given epoch (as specified by the lr policy). Args: cfg (config): configs of hyper-parameters of ADAM, includes base learning rate, betas, and weight decays. cur_epoch (float): the number of epoch of the current training stage. """ return lr_policy.get_lr_at_epoch(cfg, cur_epoch) def set_lr(optimizer, new_lr, cfg): """ Sets the optimizer lr to the specified value. Args: optimizer (optim): the optimizer using to optimize the current network. new_lr (float): the new learning rate to set. """ if cfg.TRAIN.FINETUNE: for param_group in optimizer.param_groups: if param_group['name'] == 'predictor': param_group['lr'] = new_lr[0] else: param_group['lr'] = new_lr[1] else: for param_group in optimizer.param_groups: param_group['lr'] = new_lr[0]
4,155
30.484848
101
py
STTS
STTS-main/MViT/slowfast/datasets/video_container.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import av def get_video_container(path_to_vid, multi_thread_decode=False, backend="pyav"): """ Given the path to the video, return the pyav video container. Args: path_to_vid (str): path to the video. multi_thread_decode (bool): if True, perform multi-thread decoding. backend (str): decoder backend, options include `pyav` and `torchvision`, default is `pyav`. Returns: container (container): video container. """ if backend == "torchvision": with open(path_to_vid, "rb") as fp: container = fp.read() return container elif backend == "pyav": container = av.open(path_to_vid) if multi_thread_decode: # Enable multiple threads for decoding. container.streams.video[0].thread_type = "AUTO" return container else: raise NotImplementedError("Unknown backend {}".format(backend))
1,033
33.466667
80
py
STTS
STTS-main/MViT/slowfast/datasets/transform.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import math import numpy as np # import cv2 import random import torch import torchvision.transforms.functional as F from PIL import Image from torchvision import transforms from .rand_augment import rand_augment_transform from .random_erasing import RandomErasing _pil_interpolation_to_str = { Image.NEAREST: "PIL.Image.NEAREST", Image.BILINEAR: "PIL.Image.BILINEAR", Image.BICUBIC: "PIL.Image.BICUBIC", Image.LANCZOS: "PIL.Image.LANCZOS", Image.HAMMING: "PIL.Image.HAMMING", Image.BOX: "PIL.Image.BOX", } _RANDOM_INTERPOLATION = (Image.BILINEAR, Image.BICUBIC) def _pil_interp(method): if method == "bicubic": return Image.BICUBIC elif method == "lanczos": return Image.LANCZOS elif method == "hamming": return Image.HAMMING else: return Image.BILINEAR def random_short_side_scale_jitter( images, min_size, max_size, boxes=None, inverse_uniform_sampling=False ): """ Perform a spatial short scale jittering on the given images and corresponding boxes. Args: images (tensor): images to perform scale jitter. Dimension is `num frames` x `channel` x `height` x `width`. min_size (int): the minimal size to scale the frames. max_size (int): the maximal size to scale the frames. boxes (ndarray): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. inverse_uniform_sampling (bool): if True, sample uniformly in [1 / max_scale, 1 / min_scale] and take a reciprocal to get the scale. If False, take a uniform sample from [min_scale, max_scale]. Returns: (tensor): the scaled images with dimension of `num frames` x `channel` x `new height` x `new width`. (ndarray or None): the scaled boxes with dimension of `num boxes` x 4. """ if inverse_uniform_sampling: size = int( round(1.0 / np.random.uniform(1.0 / max_size, 1.0 / min_size)) ) else: size = int(round(np.random.uniform(min_size, max_size))) height = images.shape[2] width = images.shape[3] if (width <= height and width == size) or ( height <= width and height == size ): return images, boxes new_width = size new_height = size if width < height: new_height = int(math.floor((float(height) / width) * size)) if boxes is not None: boxes = boxes * float(new_height) / height else: new_width = int(math.floor((float(width) / height) * size)) if boxes is not None: boxes = boxes * float(new_width) / width return ( torch.nn.functional.interpolate( images, size=(new_height, new_width), mode="bilinear", align_corners=False, ), boxes, ) def crop_boxes(boxes, x_offset, y_offset): """ Peform crop on the bounding boxes given the offsets. Args: boxes (ndarray or None): bounding boxes to peform crop. The dimension is `num boxes` x 4. x_offset (int): cropping offset in the x axis. y_offset (int): cropping offset in the y axis. Returns: cropped_boxes (ndarray or None): the cropped boxes with dimension of `num boxes` x 4. """ cropped_boxes = boxes.copy() cropped_boxes[:, [0, 2]] = boxes[:, [0, 2]] - x_offset cropped_boxes[:, [1, 3]] = boxes[:, [1, 3]] - y_offset return cropped_boxes def random_crop(images, size, boxes=None): """ Perform random spatial crop on the given images and corresponding boxes. Args: images (tensor): images to perform random crop. The dimension is `num frames` x `channel` x `height` x `width`. size (int): the size of height and width to crop on the image. boxes (ndarray or None): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. Returns: cropped (tensor): cropped images with dimension of `num frames` x `channel` x `size` x `size`. cropped_boxes (ndarray or None): the cropped boxes with dimension of `num boxes` x 4. """ if images.shape[2] == size and images.shape[3] == size: return images height = images.shape[2] width = images.shape[3] y_offset = 0 if height > size: y_offset = int(np.random.randint(0, height - size)) x_offset = 0 if width > size: x_offset = int(np.random.randint(0, width - size)) cropped = images[ :, :, y_offset : y_offset + size, x_offset : x_offset + size ] cropped_boxes = ( crop_boxes(boxes, x_offset, y_offset) if boxes is not None else None ) return cropped, cropped_boxes def horizontal_flip(prob, images, boxes=None): """ Perform horizontal flip on the given images and corresponding boxes. Args: prob (float): probility to flip the images. images (tensor): images to perform horizontal flip, the dimension is `num frames` x `channel` x `height` x `width`. boxes (ndarray or None): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. Returns: images (tensor): images with dimension of `num frames` x `channel` x `height` x `width`. flipped_boxes (ndarray or None): the flipped boxes with dimension of `num boxes` x 4. """ if boxes is None: flipped_boxes = None else: flipped_boxes = boxes.copy() if np.random.uniform() < prob: images = images.flip((-1)) if len(images.shape) == 3: width = images.shape[2] elif len(images.shape) == 4: width = images.shape[3] else: raise NotImplementedError("Dimension does not supported") if boxes is not None: flipped_boxes[:, [0, 2]] = width - boxes[:, [2, 0]] - 1 return images, flipped_boxes def uniform_crop(images, size, spatial_idx, boxes=None, scale_size=None): """ Perform uniform spatial sampling on the images and corresponding boxes. Args: images (tensor): images to perform uniform crop. The dimension is `num frames` x `channel` x `height` x `width`. size (int): size of height and weight to crop the images. spatial_idx (int): 0, 1, or 2 for left, center, and right crop if width is larger than height. Or 0, 1, or 2 for top, center, and bottom crop if height is larger than width. boxes (ndarray or None): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. scale_size (int): optinal. If not None, resize the images to scale_size before performing any crop. Returns: cropped (tensor): images with dimension of `num frames` x `channel` x `size` x `size`. cropped_boxes (ndarray or None): the cropped boxes with dimension of `num boxes` x 4. """ assert spatial_idx in [0, 1, 2] ndim = len(images.shape) if ndim == 3: images = images.unsqueeze(0) height = images.shape[2] width = images.shape[3] if scale_size is not None: if width <= height: width, height = scale_size, int(height / width * scale_size) else: width, height = int(width / height * scale_size), scale_size images = torch.nn.functional.interpolate( images, size=(height, width), mode="bilinear", align_corners=False, ) y_offset = int(math.ceil((height - size) / 2)) x_offset = int(math.ceil((width - size) / 2)) if height > width: if spatial_idx == 0: y_offset = 0 elif spatial_idx == 2: y_offset = height - size else: if spatial_idx == 0: x_offset = 0 elif spatial_idx == 2: x_offset = width - size cropped = images[ :, :, y_offset : y_offset + size, x_offset : x_offset + size ] cropped_boxes = ( crop_boxes(boxes, x_offset, y_offset) if boxes is not None else None ) if ndim == 3: cropped = cropped.squeeze(0) return cropped, cropped_boxes def clip_boxes_to_image(boxes, height, width): """ Clip an array of boxes to an image with the given height and width. Args: boxes (ndarray): bounding boxes to perform clipping. Dimension is `num boxes` x 4. height (int): given image height. width (int): given image width. Returns: clipped_boxes (ndarray): the clipped boxes with dimension of `num boxes` x 4. """ clipped_boxes = boxes.copy() clipped_boxes[:, [0, 2]] = np.minimum( width - 1.0, np.maximum(0.0, boxes[:, [0, 2]]) ) clipped_boxes[:, [1, 3]] = np.minimum( height - 1.0, np.maximum(0.0, boxes[:, [1, 3]]) ) return clipped_boxes def blend(images1, images2, alpha): """ Blend two images with a given weight alpha. Args: images1 (tensor): the first images to be blended, the dimension is `num frames` x `channel` x `height` x `width`. images2 (tensor): the second images to be blended, the dimension is `num frames` x `channel` x `height` x `width`. alpha (float): the blending weight. Returns: (tensor): blended images, the dimension is `num frames` x `channel` x `height` x `width`. """ return images1 * alpha + images2 * (1 - alpha) def grayscale(images): """ Get the grayscale for the input images. The channels of images should be in order BGR. Args: images (tensor): the input images for getting grayscale. Dimension is `num frames` x `channel` x `height` x `width`. Returns: img_gray (tensor): blended images, the dimension is `num frames` x `channel` x `height` x `width`. """ # R -> 0.299, G -> 0.587, B -> 0.114. img_gray = torch.tensor(images) gray_channel = ( 0.299 * images[:, 2] + 0.587 * images[:, 1] + 0.114 * images[:, 0] ) img_gray[:, 0] = gray_channel img_gray[:, 1] = gray_channel img_gray[:, 2] = gray_channel return img_gray def color_jitter(images, img_brightness=0, img_contrast=0, img_saturation=0): """ Perfrom a color jittering on the input images. The channels of images should be in order BGR. Args: images (tensor): images to perform color jitter. Dimension is `num frames` x `channel` x `height` x `width`. img_brightness (float): jitter ratio for brightness. img_contrast (float): jitter ratio for contrast. img_saturation (float): jitter ratio for saturation. Returns: images (tensor): the jittered images, the dimension is `num frames` x `channel` x `height` x `width`. """ jitter = [] if img_brightness != 0: jitter.append("brightness") if img_contrast != 0: jitter.append("contrast") if img_saturation != 0: jitter.append("saturation") if len(jitter) > 0: order = np.random.permutation(np.arange(len(jitter))) for idx in range(0, len(jitter)): if jitter[order[idx]] == "brightness": images = brightness_jitter(img_brightness, images) elif jitter[order[idx]] == "contrast": images = contrast_jitter(img_contrast, images) elif jitter[order[idx]] == "saturation": images = saturation_jitter(img_saturation, images) return images def brightness_jitter(var, images): """ Perfrom brightness jittering on the input images. The channels of images should be in order BGR. Args: var (float): jitter ratio for brightness. images (tensor): images to perform color jitter. Dimension is `num frames` x `channel` x `height` x `width`. Returns: images (tensor): the jittered images, the dimension is `num frames` x `channel` x `height` x `width`. """ alpha = 1.0 + np.random.uniform(-var, var) img_bright = torch.zeros(images.shape) images = blend(images, img_bright, alpha) return images def contrast_jitter(var, images): """ Perfrom contrast jittering on the input images. The channels of images should be in order BGR. Args: var (float): jitter ratio for contrast. images (tensor): images to perform color jitter. Dimension is `num frames` x `channel` x `height` x `width`. Returns: images (tensor): the jittered images, the dimension is `num frames` x `channel` x `height` x `width`. """ alpha = 1.0 + np.random.uniform(-var, var) img_gray = grayscale(images) img_gray[:] = torch.mean(img_gray, dim=(1, 2, 3), keepdim=True) images = blend(images, img_gray, alpha) return images def saturation_jitter(var, images): """ Perfrom saturation jittering on the input images. The channels of images should be in order BGR. Args: var (float): jitter ratio for saturation. images (tensor): images to perform color jitter. Dimension is `num frames` x `channel` x `height` x `width`. Returns: images (tensor): the jittered images, the dimension is `num frames` x `channel` x `height` x `width`. """ alpha = 1.0 + np.random.uniform(-var, var) img_gray = grayscale(images) images = blend(images, img_gray, alpha) return images def lighting_jitter(images, alphastd, eigval, eigvec): """ Perform AlexNet-style PCA jitter on the given images. Args: images (tensor): images to perform lighting jitter. Dimension is `num frames` x `channel` x `height` x `width`. alphastd (float): jitter ratio for PCA jitter. eigval (list): eigenvalues for PCA jitter. eigvec (list[list]): eigenvectors for PCA jitter. Returns: out_images (tensor): the jittered images, the dimension is `num frames` x `channel` x `height` x `width`. """ if alphastd == 0: return images # generate alpha1, alpha2, alpha3. alpha = np.random.normal(0, alphastd, size=(1, 3)) eig_vec = np.array(eigvec) eig_val = np.reshape(eigval, (1, 3)) rgb = np.sum( eig_vec * np.repeat(alpha, 3, axis=0) * np.repeat(eig_val, 3, axis=0), axis=1, ) out_images = torch.zeros_like(images) if len(images.shape) == 3: # C H W channel_dim = 0 elif len(images.shape) == 4: # T C H W channel_dim = 1 else: raise NotImplementedError(f"Unsupported dimension {len(images.shape)}") for idx in range(images.shape[channel_dim]): # C H W if len(images.shape) == 3: out_images[idx] = images[idx] + rgb[2 - idx] # T C H W elif len(images.shape) == 4: out_images[:, idx] = images[:, idx] + rgb[2 - idx] else: raise NotImplementedError( f"Unsupported dimension {len(images.shape)}" ) return out_images def color_normalization(images, mean, stddev): """ Perform color nomration on the given images. Args: images (tensor): images to perform color normalization. Dimension is `num frames` x `channel` x `height` x `width`. mean (list): mean values for normalization. stddev (list): standard deviations for normalization. Returns: out_images (tensor): the noramlized images, the dimension is `num frames` x `channel` x `height` x `width`. """ if len(images.shape) == 3: assert ( len(mean) == images.shape[0] ), "channel mean not computed properly" assert ( len(stddev) == images.shape[0] ), "channel stddev not computed properly" elif len(images.shape) == 4: assert ( len(mean) == images.shape[1] ), "channel mean not computed properly" assert ( len(stddev) == images.shape[1] ), "channel stddev not computed properly" else: raise NotImplementedError(f"Unsupported dimension {len(images.shape)}") out_images = torch.zeros_like(images) for idx in range(len(mean)): # C H W if len(images.shape) == 3: out_images[idx] = (images[idx] - mean[idx]) / stddev[idx] elif len(images.shape) == 4: out_images[:, idx] = (images[:, idx] - mean[idx]) / stddev[idx] else: raise NotImplementedError( f"Unsupported dimension {len(images.shape)}" ) return out_images def _get_param_spatial_crop( scale, ratio, height, width, num_repeat=10, log_scale=True, switch_hw=False ): """ Given scale, ratio, height and width, return sampled coordinates of the videos. """ for _ in range(num_repeat): area = height * width target_area = random.uniform(*scale) * area if log_scale: log_ratio = (math.log(ratio[0]), math.log(ratio[1])) aspect_ratio = math.exp(random.uniform(*log_ratio)) else: aspect_ratio = random.uniform(*ratio) w = int(round(math.sqrt(target_area * aspect_ratio))) h = int(round(math.sqrt(target_area / aspect_ratio))) if np.random.uniform() < 0.5 and switch_hw: w, h = h, w if 0 < w <= width and 0 < h <= height: i = random.randint(0, height - h) j = random.randint(0, width - w) return i, j, h, w # Fallback to central crop in_ratio = float(width) / float(height) if in_ratio < min(ratio): w = width h = int(round(w / min(ratio))) elif in_ratio > max(ratio): h = height w = int(round(h * max(ratio))) else: # whole image w = width h = height i = (height - h) // 2 j = (width - w) // 2 return i, j, h, w def random_resized_crop( images, target_height, target_width, scale=(0.8, 1.0), ratio=(3.0 / 4.0, 4.0 / 3.0), ): """ Crop the given images to random size and aspect ratio. A crop of random size (default: of 0.08 to 1.0) of the original size and a random aspect ratio (default: of 3/4 to 4/3) of the original aspect ratio is made. This crop is finally resized to given size. This is popularly used to train the Inception networks. Args: images: Images to perform resizing and cropping. target_height: Desired height after cropping. target_width: Desired width after cropping. scale: Scale range of Inception-style area based random resizing. ratio: Aspect ratio range of Inception-style area based random resizing. """ height = images.shape[2] width = images.shape[3] i, j, h, w = _get_param_spatial_crop(scale, ratio, height, width) cropped = images[:, :, i : i + h, j : j + w] return torch.nn.functional.interpolate( cropped, size=(target_height, target_width), mode="bilinear", align_corners=False, ) def random_resized_crop_with_shift( images, target_height, target_width, scale=(0.8, 1.0), ratio=(3.0 / 4.0, 4.0 / 3.0), ): """ This is similar to random_resized_crop. However, it samples two different boxes (for cropping) for the first and last frame. It then linearly interpolates the two boxes for other frames. Args: images: Images to perform resizing and cropping. target_height: Desired height after cropping. target_width: Desired width after cropping. scale: Scale range of Inception-style area based random resizing. ratio: Aspect ratio range of Inception-style area based random resizing. """ t = images.shape[1] height = images.shape[2] width = images.shape[3] i, j, h, w = _get_param_spatial_crop(scale, ratio, height, width) i_, j_, h_, w_ = _get_param_spatial_crop(scale, ratio, height, width) i_s = [int(i) for i in torch.linspace(i, i_, steps=t).tolist()] j_s = [int(i) for i in torch.linspace(j, j_, steps=t).tolist()] h_s = [int(i) for i in torch.linspace(h, h_, steps=t).tolist()] w_s = [int(i) for i in torch.linspace(w, w_, steps=t).tolist()] out = torch.zeros((3, t, target_height, target_width)) for ind in range(t): out[:, ind : ind + 1, :, :] = torch.nn.functional.interpolate( images[ :, ind : ind + 1, i_s[ind] : i_s[ind] + h_s[ind], j_s[ind] : j_s[ind] + w_s[ind], ], size=(target_height, target_width), mode="bilinear", align_corners=False, ) return out def create_random_augment( input_size, auto_augment=None, interpolation="bilinear", ): """ Get video randaug transform. Args: input_size: The size of the input video in tuple. auto_augment: Parameters for randaug. An example: "rand-m7-n4-mstd0.5-inc1" (m is the magnitude and n is the number of operations to apply). interpolation: Interpolation method. """ if isinstance(input_size, tuple): img_size = input_size[-2:] else: img_size = input_size if auto_augment: assert isinstance(auto_augment, str) if isinstance(img_size, tuple): img_size_min = min(img_size) else: img_size_min = img_size aa_params = {"translate_const": int(img_size_min * 0.45)} if interpolation and interpolation != "random": aa_params["interpolation"] = _pil_interp(interpolation) if auto_augment.startswith("rand"): return transforms.Compose( [rand_augment_transform(auto_augment, aa_params)] ) raise NotImplementedError def random_sized_crop_img( im, size, jitter_scale=(0.08, 1.0), jitter_aspect=(3.0 / 4.0, 4.0 / 3.0), max_iter=10, ): """ Performs Inception-style cropping (used for training). """ assert ( len(im.shape) == 3 ), "Currently only support image for random_sized_crop" h, w = im.shape[1:3] i, j, h, w = _get_param_spatial_crop( scale=jitter_scale, ratio=jitter_aspect, height=h, width=w, num_repeat=max_iter, log_scale=False, switch_hw=True, ) cropped = im[:, i : i + h, j : j + w] return torch.nn.functional.interpolate( cropped.unsqueeze(0), size=(size, size), mode="bilinear", align_corners=False, ).squeeze(0) # The following code are modified based on timm lib, we will replace the following # contents with dependency from PyTorchVideo. # https://github.com/facebookresearch/pytorchvideo class RandomResizedCropAndInterpolation: """Crop the given PIL Image to random size and aspect ratio with random interpolation. A crop of random size (default: of 0.08 to 1.0) of the original size and a random aspect ratio (default: of 3/4 to 4/3) of the original aspect ratio is made. This crop is finally resized to given size. This is popularly used to train the Inception networks. Args: size: expected output size of each edge scale: range of size of the origin size cropped ratio: range of aspect ratio of the origin aspect ratio cropped interpolation: Default: PIL.Image.BILINEAR """ def __init__( self, size, scale=(0.08, 1.0), ratio=(3.0 / 4.0, 4.0 / 3.0), interpolation="bilinear", ): if isinstance(size, tuple): self.size = size else: self.size = (size, size) if (scale[0] > scale[1]) or (ratio[0] > ratio[1]): print("range should be of kind (min, max)") if interpolation == "random": self.interpolation = _RANDOM_INTERPOLATION else: self.interpolation = _pil_interp(interpolation) self.scale = scale self.ratio = ratio @staticmethod def get_params(img, scale, ratio): """Get parameters for ``crop`` for a random sized crop. Args: img (PIL Image): Image to be cropped. scale (tuple): range of size of the origin size cropped ratio (tuple): range of aspect ratio of the origin aspect ratio cropped Returns: tuple: params (i, j, h, w) to be passed to ``crop`` for a random sized crop. """ area = img.size[0] * img.size[1] for _ in range(10): target_area = random.uniform(*scale) * area log_ratio = (math.log(ratio[0]), math.log(ratio[1])) aspect_ratio = math.exp(random.uniform(*log_ratio)) w = int(round(math.sqrt(target_area * aspect_ratio))) h = int(round(math.sqrt(target_area / aspect_ratio))) if w <= img.size[0] and h <= img.size[1]: i = random.randint(0, img.size[1] - h) j = random.randint(0, img.size[0] - w) return i, j, h, w # Fallback to central crop in_ratio = img.size[0] / img.size[1] if in_ratio < min(ratio): w = img.size[0] h = int(round(w / min(ratio))) elif in_ratio > max(ratio): h = img.size[1] w = int(round(h * max(ratio))) else: # whole image w = img.size[0] h = img.size[1] i = (img.size[1] - h) // 2 j = (img.size[0] - w) // 2 return i, j, h, w def __call__(self, img): """ Args: img (PIL Image): Image to be cropped and resized. Returns: PIL Image: Randomly cropped and resized image. """ i, j, h, w = self.get_params(img, self.scale, self.ratio) if isinstance(self.interpolation, (tuple, list)): interpolation = random.choice(self.interpolation) else: interpolation = self.interpolation return F.resized_crop(img, i, j, h, w, self.size, interpolation) def __repr__(self): if isinstance(self.interpolation, (tuple, list)): interpolate_str = " ".join( [_pil_interpolation_to_str[x] for x in self.interpolation] ) else: interpolate_str = _pil_interpolation_to_str[self.interpolation] format_string = self.__class__.__name__ + "(size={0}".format(self.size) format_string += ", scale={0}".format( tuple(round(s, 4) for s in self.scale) ) format_string += ", ratio={0}".format( tuple(round(r, 4) for r in self.ratio) ) format_string += ", interpolation={0})".format(interpolate_str) return format_string def transforms_imagenet_train( img_size=224, scale=None, ratio=None, hflip=0.5, vflip=0.0, color_jitter=0.4, auto_augment=None, interpolation="random", use_prefetcher=False, mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225), re_prob=0.0, re_mode="const", re_count=1, re_num_splits=0, separate=False, ): """ If separate==True, the transforms are returned as a tuple of 3 separate transforms for use in a mixing dataset that passes * all data through the first (primary) transform, called the 'clean' data * a portion of the data through the secondary transform * normalizes and converts the branches above with the third, final transform """ if isinstance(img_size, tuple): img_size = img_size[-2:] else: img_size = img_size scale = tuple(scale or (0.08, 1.0)) # default imagenet scale range ratio = tuple( ratio or (3.0 / 4.0, 4.0 / 3.0) ) # default imagenet ratio range primary_tfl = [ RandomResizedCropAndInterpolation( img_size, scale=scale, ratio=ratio, interpolation=interpolation ) ] if hflip > 0.0: primary_tfl += [transforms.RandomHorizontalFlip(p=hflip)] if vflip > 0.0: primary_tfl += [transforms.RandomVerticalFlip(p=vflip)] secondary_tfl = [] if auto_augment: assert isinstance(auto_augment, str) if isinstance(img_size, tuple): img_size_min = min(img_size) else: img_size_min = img_size aa_params = dict( translate_const=int(img_size_min * 0.45), img_mean=tuple([min(255, round(255 * x)) for x in mean]), ) if interpolation and interpolation != "random": aa_params["interpolation"] = _pil_interp(interpolation) if auto_augment.startswith("rand"): secondary_tfl += [rand_augment_transform(auto_augment, aa_params)] elif auto_augment.startswith("augmix"): raise NotImplementedError("Augmix not implemented") else: raise NotImplementedError("Auto aug not implemented") elif color_jitter is not None: # color jitter is enabled when not using AA if isinstance(color_jitter, (list, tuple)): # color jitter should be a 3-tuple/list if spec brightness/contrast/saturation # or 4 if also augmenting hue assert len(color_jitter) in (3, 4) else: # if it's a scalar, duplicate for brightness, contrast, and saturation, no hue color_jitter = (float(color_jitter),) * 3 secondary_tfl += [transforms.ColorJitter(*color_jitter)] final_tfl = [] final_tfl += [ transforms.ToTensor(), transforms.Normalize(mean=torch.tensor(mean), std=torch.tensor(std)), ] if re_prob > 0.0: final_tfl.append( RandomErasing( re_prob, mode=re_mode, max_count=re_count, num_splits=re_num_splits, device="cpu", cube=False, ) ) if separate: return ( transforms.Compose(primary_tfl), transforms.Compose(secondary_tfl), transforms.Compose(final_tfl), ) else: return transforms.Compose(primary_tfl + secondary_tfl + final_tfl)
30,520
33.101676
90
py
STTS
STTS-main/MViT/slowfast/datasets/utils.py
#!/usr/bin/env python3 import logging import numpy as np import os import random import time from collections import defaultdict import cv2 import torch from torch.utils.data.distributed import DistributedSampler from slowfast.utils.env import pathmgr from . import transform as transform logger = logging.getLogger(__name__) def retry_load_images(image_paths, retry=10, backend="pytorch"): """ This function is to load images with support of retrying for failed load. Args: image_paths (list): paths of images needed to be loaded. retry (int, optional): maximum time of loading retrying. Defaults to 10. backend (str): `pytorch` or `cv2`. Returns: imgs (list): list of loaded images. """ for i in range(retry): imgs = [] for image_path in image_paths: with pathmgr.open(image_path, "rb") as f: img_str = np.frombuffer(f.read(), np.uint8) img = cv2.imdecode(img_str, flags=cv2.IMREAD_COLOR) imgs.append(img) if all(img is not None for img in imgs): if backend == "pytorch": imgs = torch.as_tensor(np.stack(imgs)) return imgs else: logger.warn("Reading failed. Will retry.") time.sleep(1.0) if i == retry - 1: raise Exception("Failed to load images {}".format(image_paths)) def get_sequence(center_idx, half_len, sample_rate, num_frames): """ Sample frames among the corresponding clip. Args: center_idx (int): center frame idx for current clip half_len (int): half of the clip length sample_rate (int): sampling rate for sampling frames inside of the clip num_frames (int): number of expected sampled frames Returns: seq (list): list of indexes of sampled frames in this clip. """ seq = list(range(center_idx - half_len, center_idx + half_len, sample_rate)) for seq_idx in range(len(seq)): if seq[seq_idx] < 0: seq[seq_idx] = 0 elif seq[seq_idx] >= num_frames: seq[seq_idx] = num_frames - 1 return seq def pack_pathway_output(cfg, frames): """ Prepare output as a list of tensors. Each tensor corresponding to a unique pathway. Args: frames (tensor): frames of images sampled from the video. The dimension is `channel` x `num frames` x `height` x `width`. Returns: frame_list (list): list of tensors with the dimension of `channel` x `num frames` x `height` x `width`. """ if cfg.DATA.REVERSE_INPUT_CHANNEL: frames = frames[[2, 1, 0], :, :, :] if cfg.MODEL.ARCH in cfg.MODEL.SINGLE_PATHWAY_ARCH: frame_list = [frames] elif cfg.MODEL.ARCH in cfg.MODEL.MULTI_PATHWAY_ARCH: fast_pathway = frames # Perform temporal sampling from the fast pathway. slow_pathway = torch.index_select( frames, 1, torch.linspace( 0, frames.shape[1] - 1, frames.shape[1] // cfg.SLOWFAST.ALPHA ).long(), ) frame_list = [slow_pathway, fast_pathway] else: raise NotImplementedError( "Model arch {} is not in {}".format( cfg.MODEL.ARCH, cfg.MODEL.SINGLE_PATHWAY_ARCH + cfg.MODEL.MULTI_PATHWAY_ARCH, ) ) return frame_list def spatial_sampling( frames, spatial_idx=-1, min_scale=256, max_scale=320, crop_size=224, random_horizontal_flip=True, inverse_uniform_sampling=False, aspect_ratio=None, scale=None, motion_shift=False, ): """ Perform spatial sampling on the given video frames. If spatial_idx is -1, perform random scale, random crop, and random flip on the given frames. If spatial_idx is 0, 1, or 2, perform spatial uniform sampling with the given spatial_idx. Args: frames (tensor): frames of images sampled from the video. The dimension is `num frames` x `height` x `width` x `channel`. spatial_idx (int): if -1, perform random spatial sampling. If 0, 1, or 2, perform left, center, right crop if width is larger than height, and perform top, center, buttom crop if height is larger than width. min_scale (int): the minimal size of scaling. max_scale (int): the maximal size of scaling. crop_size (int): the size of height and width used to crop the frames. inverse_uniform_sampling (bool): if True, sample uniformly in [1 / max_scale, 1 / min_scale] and take a reciprocal to get the scale. If False, take a uniform sample from [min_scale, max_scale]. aspect_ratio (list): Aspect ratio range for resizing. scale (list): Scale range for resizing. motion_shift (bool): Whether to apply motion shift for resizing. Returns: frames (tensor): spatially sampled frames. """ assert spatial_idx in [-1, 0, 1, 2] if spatial_idx == -1: if aspect_ratio is None and scale is None: frames, _ = transform.random_short_side_scale_jitter( images=frames, min_size=min_scale, max_size=max_scale, inverse_uniform_sampling=inverse_uniform_sampling, ) frames, _ = transform.random_crop(frames, crop_size) else: transform_func = ( transform.random_resized_crop_with_shift if motion_shift else transform.random_resized_crop ) frames = transform_func( images=frames, target_height=crop_size, target_width=crop_size, scale=scale, ratio=aspect_ratio, ) if random_horizontal_flip: frames, _ = transform.horizontal_flip(0.5, frames) else: # The testing is deterministic and no jitter should be performed. # min_scale, max_scale, and crop_size are expect to be the same. assert len({min_scale, max_scale}) == 1 frames, _ = transform.random_short_side_scale_jitter( frames, min_scale, max_scale ) frames, _ = transform.uniform_crop(frames, crop_size, spatial_idx) return frames def as_binary_vector(labels, num_classes): """ Construct binary label vector given a list of label indices. Args: labels (list): The input label list. num_classes (int): Number of classes of the label vector. Returns: labels (numpy array): the resulting binary vector. """ label_arr = np.zeros((num_classes,)) for lbl in set(labels): label_arr[lbl] = 1.0 return label_arr def aggregate_labels(label_list): """ Join a list of label list. Args: labels (list): The input label list. Returns: labels (list): The joint list of all lists in input. """ all_labels = [] for labels in label_list: for l in labels: all_labels.append(l) return list(set(all_labels)) def convert_to_video_level_labels(labels): """ Aggregate annotations from all frames of a video to form video-level labels. Args: labels (list): The input label list. Returns: labels (list): Same as input, but with each label replaced by a video-level one. """ for video_id in range(len(labels)): video_level_labels = aggregate_labels(labels[video_id]) for i in range(len(labels[video_id])): labels[video_id][i] = video_level_labels return labels def load_image_lists(frame_list_file, prefix="", return_list=False): """ Load image paths and labels from a "frame list". Each line of the frame list contains: `original_vido_id video_id frame_id path labels` Args: frame_list_file (string): path to the frame list. prefix (str): the prefix for the path. return_list (bool): if True, return a list. If False, return a dict. Returns: image_paths (list or dict): list of list containing path to each frame. If return_list is False, then return in a dict form. labels (list or dict): list of list containing label of each frame. If return_list is False, then return in a dict form. """ image_paths = defaultdict(list) labels = defaultdict(list) with pathmgr.open(frame_list_file, "r") as f: assert f.readline().startswith("original_vido_id") for line in f: row = line.split() # original_vido_id video_id frame_id path labels assert len(row) == 5 video_name = row[0] if prefix == "": path = row[3] else: path = os.path.join(prefix, row[3]) image_paths[video_name].append(path) frame_labels = row[-1].replace('"', "") if frame_labels != "": labels[video_name].append( [int(x) for x in frame_labels.split(",")] ) else: labels[video_name].append([]) if return_list: keys = image_paths.keys() image_paths = [image_paths[key] for key in keys] labels = [labels[key] for key in keys] return image_paths, labels return dict(image_paths), dict(labels) def tensor_normalize(tensor, mean, std): """ Normalize a given tensor by subtracting the mean and dividing the std. Args: tensor (tensor): tensor to normalize. mean (tensor or list): mean value to subtract. std (tensor or list): std to divide. """ if tensor.dtype == torch.uint8: tensor = tensor.float() tensor = tensor / 255.0 if type(mean) == list: mean = torch.tensor(mean) if type(std) == list: std = torch.tensor(std) tensor = tensor - mean tensor = tensor / std return tensor def get_random_sampling_rate(long_cycle_sampling_rate, sampling_rate): """ When multigrid training uses a fewer number of frames, we randomly increase the sampling rate so that some clips cover the original span. """ if long_cycle_sampling_rate > 0: assert long_cycle_sampling_rate >= sampling_rate return random.randint(sampling_rate, long_cycle_sampling_rate) else: return sampling_rate def revert_tensor_normalize(tensor, mean, std): """ Revert normalization for a given tensor by multiplying by the std and adding the mean. Args: tensor (tensor): tensor to revert normalization. mean (tensor or list): mean value to add. std (tensor or list): std to multiply. """ if type(mean) == list: mean = torch.tensor(mean) if type(std) == list: std = torch.tensor(std) tensor = tensor * std tensor = tensor + mean return tensor def create_sampler(dataset, shuffle, cfg): """ Create sampler for the given dataset. Args: dataset (torch.utils.data.Dataset): the given dataset. shuffle (bool): set to ``True`` to have the data reshuffled at every epoch. cfg (CfgNode): configs. Details can be found in slowfast/config/defaults.py Returns: sampler (Sampler): the created sampler. """ sampler = DistributedSampler(dataset) if cfg.NUM_GPUS > 1 else None return sampler def loader_worker_init_fn(dataset): """ Create init function passed to pytorch data loader. Args: dataset (torch.utils.data.Dataset): the given dataset. """ return None
11,739
32.638968
90
py
STTS
STTS-main/MViT/slowfast/datasets/mixup.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """ This implementation is based on https://github.com/rwightman/pytorch-image-models/blob/master/timm/data/mixup.py, published under an Apache License 2.0. COMMENT FROM ORIGINAL: Mixup and Cutmix Papers: mixup: Beyond Empirical Risk Minimization (https://arxiv.org/abs/1710.09412) CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features (https://arxiv.org/abs/1905.04899) # NOQA Code Reference: CutMix: https://github.com/clovaai/CutMix-PyTorch Hacked together by / Copyright 2020 Ross Wightman """ import numpy as np import torch def convert_to_one_hot(targets, num_classes, on_value=1.0, off_value=0.0): """ This function converts target class indices to one-hot vectors, given the number of classes. Args: targets (loader): Class labels. num_classes (int): Total number of classes. on_value (float): Target Value for ground truth class. off_value (float): Target Value for other classes.This value is used for label smoothing. """ targets = targets.long().view(-1, 1) return torch.full( (targets.size()[0], num_classes), off_value, device=targets.device ).scatter_(1, targets, on_value) def mixup_target(target, num_classes, lam=1.0, smoothing=0.0): """ This function converts target class indices to one-hot vectors, given the number of classes. Args: targets (loader): Class labels. num_classes (int): Total number of classes. lam (float): lamba value for mixup/cutmix. smoothing (float): Label smoothing value. """ off_value = smoothing / num_classes on_value = 1.0 - smoothing + off_value target1 = convert_to_one_hot( target, num_classes, on_value=on_value, off_value=off_value, ) target2 = convert_to_one_hot( target.flip(0), num_classes, on_value=on_value, off_value=off_value, ) return target1 * lam + target2 * (1.0 - lam) def rand_bbox(img_shape, lam, margin=0.0, count=None): """ Generates a random square bbox based on lambda value. Args: img_shape (tuple): Image shape as tuple lam (float): Cutmix lambda value margin (float): Percentage of bbox dimension to enforce as margin (reduce amount of box outside image) count (int): Number of bbox to generate """ ratio = np.sqrt(1 - lam) img_h, img_w = img_shape[-2:] cut_h, cut_w = int(img_h * ratio), int(img_w * ratio) margin_y, margin_x = int(margin * cut_h), int(margin * cut_w) cy = np.random.randint(0 + margin_y, img_h - margin_y, size=count) cx = np.random.randint(0 + margin_x, img_w - margin_x, size=count) yl = np.clip(cy - cut_h // 2, 0, img_h) yh = np.clip(cy + cut_h // 2, 0, img_h) xl = np.clip(cx - cut_w // 2, 0, img_w) xh = np.clip(cx + cut_w // 2, 0, img_w) return yl, yh, xl, xh def get_cutmix_bbox(img_shape, lam, correct_lam=True, count=None): """ Generates the box coordinates for cutmix. Args: img_shape (tuple): Image shape as tuple lam (float): Cutmix lambda value correct_lam (bool): Apply lambda correction when cutmix bbox clipped by image borders. count (int): Number of bbox to generate """ yl, yu, xl, xu = rand_bbox(img_shape, lam, count=count) if correct_lam: bbox_area = (yu - yl) * (xu - xl) lam = 1.0 - bbox_area / float(img_shape[-2] * img_shape[-1]) return (yl, yu, xl, xu), lam class MixUp: """ Apply mixup and/or cutmix for videos at batch level. mixup: Beyond Empirical Risk Minimization (https://arxiv.org/abs/1710.09412) CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features (https://arxiv.org/abs/1905.04899) """ def __init__( self, mixup_alpha=1.0, cutmix_alpha=0.0, mix_prob=1.0, switch_prob=0.5, correct_lam=True, label_smoothing=0.1, num_classes=1000, ): """ Args: mixup_alpha (float): Mixup alpha value. cutmix_alpha (float): Cutmix alpha value. mix_prob (float): Probability of applying mixup or cutmix. switch_prob (float): Probability of switching to cutmix instead of mixup when both are active. correct_lam (bool): Apply lambda correction when cutmix bbox clipped by image borders. label_smoothing (float): Apply label smoothing to the mixed target tensor. If label_smoothing is not used, set it to 0. num_classes (int): Number of classes for target. """ self.mixup_alpha = mixup_alpha self.cutmix_alpha = cutmix_alpha self.mix_prob = mix_prob self.switch_prob = switch_prob self.label_smoothing = label_smoothing self.num_classes = num_classes self.correct_lam = correct_lam def _get_mixup_params(self): lam = 1.0 use_cutmix = False if np.random.rand() < self.mix_prob: if self.mixup_alpha > 0.0 and self.cutmix_alpha > 0.0: use_cutmix = np.random.rand() < self.switch_prob lam_mix = ( np.random.beta(self.cutmix_alpha, self.cutmix_alpha) if use_cutmix else np.random.beta(self.mixup_alpha, self.mixup_alpha) ) elif self.mixup_alpha > 0.0: lam_mix = np.random.beta(self.mixup_alpha, self.mixup_alpha) elif self.cutmix_alpha > 0.0: use_cutmix = True lam_mix = np.random.beta(self.cutmix_alpha, self.cutmix_alpha) lam = float(lam_mix) return lam, use_cutmix def _mix_batch(self, x): lam, use_cutmix = self._get_mixup_params() if lam == 1.0: return 1.0 if use_cutmix: (yl, yh, xl, xh), lam = get_cutmix_bbox( x.shape, lam, correct_lam=self.correct_lam, ) x[..., yl:yh, xl:xh] = x.flip(0)[..., yl:yh, xl:xh] else: x_flipped = x.flip(0).mul_(1.0 - lam) x.mul_(lam).add_(x_flipped) return lam def __call__(self, x, target): assert len(x) > 1, "Batch size should be greater than 1 for mixup." lam = self._mix_batch(x) target = mixup_target( target, self.num_classes, lam, self.label_smoothing ) return x, target
6,659
34.052632
127
py
STTS
STTS-main/MViT/slowfast/datasets/ava_dataset.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import logging import numpy as np import torch from . import ava_helper as ava_helper from . import cv2_transform as cv2_transform from . import transform as transform from . import utils as utils from .build import DATASET_REGISTRY logger = logging.getLogger(__name__) @DATASET_REGISTRY.register() class Ava(torch.utils.data.Dataset): """ AVA Dataset """ def __init__(self, cfg, split): self.cfg = cfg self._split = split self._sample_rate = cfg.DATA.SAMPLING_RATE self._video_length = cfg.DATA.NUM_FRAMES self._seq_len = self._video_length * self._sample_rate self._num_classes = cfg.MODEL.NUM_CLASSES # Augmentation params. self._data_mean = cfg.DATA.MEAN self._data_std = cfg.DATA.STD self._use_bgr = cfg.AVA.BGR self.random_horizontal_flip = cfg.DATA.RANDOM_FLIP if self._split == "train": self._crop_size = cfg.DATA.TRAIN_CROP_SIZE self._jitter_min_scale = cfg.DATA.TRAIN_JITTER_SCALES[0] self._jitter_max_scale = cfg.DATA.TRAIN_JITTER_SCALES[1] self._use_color_augmentation = cfg.AVA.TRAIN_USE_COLOR_AUGMENTATION self._pca_jitter_only = cfg.AVA.TRAIN_PCA_JITTER_ONLY self._pca_eigval = cfg.DATA.TRAIN_PCA_EIGVAL self._pca_eigvec = cfg.DATA.TRAIN_PCA_EIGVEC else: self._crop_size = cfg.DATA.TEST_CROP_SIZE self._test_force_flip = cfg.AVA.TEST_FORCE_FLIP self._load_data(cfg) def _load_data(self, cfg): """ Load frame paths and annotations from files Args: cfg (CfgNode): config """ # Loading frame paths. ( self._image_paths, self._video_idx_to_name, ) = ava_helper.load_image_lists(cfg, is_train=(self._split == "train")) # Loading annotations for boxes and labels. boxes_and_labels = ava_helper.load_boxes_and_labels( cfg, mode=self._split ) assert len(boxes_and_labels) == len(self._image_paths) boxes_and_labels = [ boxes_and_labels[self._video_idx_to_name[i]] for i in range(len(self._image_paths)) ] # Get indices of keyframes and corresponding boxes and labels. ( self._keyframe_indices, self._keyframe_boxes_and_labels, ) = ava_helper.get_keyframe_data(boxes_and_labels) # Calculate the number of used boxes. self._num_boxes_used = ava_helper.get_num_boxes_used( self._keyframe_indices, self._keyframe_boxes_and_labels ) self.print_summary() def print_summary(self): logger.info("=== AVA dataset summary ===") logger.info("Split: {}".format(self._split)) logger.info("Number of videos: {}".format(len(self._image_paths))) total_frames = sum( len(video_img_paths) for video_img_paths in self._image_paths ) logger.info("Number of frames: {}".format(total_frames)) logger.info("Number of key frames: {}".format(len(self))) logger.info("Number of boxes: {}.".format(self._num_boxes_used)) def __len__(self): """ Returns: (int): the number of videos in the dataset. """ return self.num_videos @property def num_videos(self): """ Returns: (int): the number of videos in the dataset. """ return len(self._keyframe_indices) def _images_and_boxes_preprocessing_cv2(self, imgs, boxes): """ This function performs preprocessing for the input images and corresponding boxes for one clip with opencv as backend. Args: imgs (tensor): the images. boxes (ndarray): the boxes for the current clip. Returns: imgs (tensor): list of preprocessed images. boxes (ndarray): preprocessed boxes. """ height, width, _ = imgs[0].shape boxes[:, [0, 2]] *= width boxes[:, [1, 3]] *= height boxes = cv2_transform.clip_boxes_to_image(boxes, height, width) # `transform.py` is list of np.array. However, for AVA, we only have # one np.array. boxes = [boxes] # The image now is in HWC, BGR format. if self._split == "train": # "train" imgs, boxes = cv2_transform.random_short_side_scale_jitter_list( imgs, min_size=self._jitter_min_scale, max_size=self._jitter_max_scale, boxes=boxes, ) imgs, boxes = cv2_transform.random_crop_list( imgs, self._crop_size, order="HWC", boxes=boxes ) if self.random_horizontal_flip: # random flip imgs, boxes = cv2_transform.horizontal_flip_list( 0.5, imgs, order="HWC", boxes=boxes ) elif self._split == "val": # Short side to test_scale. Non-local and STRG uses 256. imgs = [cv2_transform.scale(self._crop_size, img) for img in imgs] boxes = [ cv2_transform.scale_boxes( self._crop_size, boxes[0], height, width ) ] imgs, boxes = cv2_transform.spatial_shift_crop_list( self._crop_size, imgs, 1, boxes=boxes ) if self._test_force_flip: imgs, boxes = cv2_transform.horizontal_flip_list( 1, imgs, order="HWC", boxes=boxes ) elif self._split == "test": # Short side to test_scale. Non-local and STRG uses 256. imgs = [cv2_transform.scale(self._crop_size, img) for img in imgs] boxes = [ cv2_transform.scale_boxes( self._crop_size, boxes[0], height, width ) ] if self._test_force_flip: imgs, boxes = cv2_transform.horizontal_flip_list( 1, imgs, order="HWC", boxes=boxes ) else: raise NotImplementedError( "Unsupported split mode {}".format(self._split) ) # Convert image to CHW keeping BGR order. imgs = [cv2_transform.HWC2CHW(img) for img in imgs] # Image [0, 255] -> [0, 1]. imgs = [img / 255.0 for img in imgs] imgs = [ np.ascontiguousarray( # img.reshape((3, self._crop_size, self._crop_size)) img.reshape((3, imgs[0].shape[1], imgs[0].shape[2])) ).astype(np.float32) for img in imgs ] # Do color augmentation (after divided by 255.0). if self._split == "train" and self._use_color_augmentation: if not self._pca_jitter_only: imgs = cv2_transform.color_jitter_list( imgs, img_brightness=0.4, img_contrast=0.4, img_saturation=0.4, ) imgs = cv2_transform.lighting_list( imgs, alphastd=0.1, eigval=np.array(self._pca_eigval).astype(np.float32), eigvec=np.array(self._pca_eigvec).astype(np.float32), ) # Normalize images by mean and std. imgs = [ cv2_transform.color_normalization( img, np.array(self._data_mean, dtype=np.float32), np.array(self._data_std, dtype=np.float32), ) for img in imgs ] # Concat list of images to single ndarray. imgs = np.concatenate( [np.expand_dims(img, axis=1) for img in imgs], axis=1 ) if not self._use_bgr: # Convert image format from BGR to RGB. imgs = imgs[::-1, ...] imgs = np.ascontiguousarray(imgs) imgs = torch.from_numpy(imgs) boxes = cv2_transform.clip_boxes_to_image( boxes[0], imgs[0].shape[1], imgs[0].shape[2] ) return imgs, boxes def _images_and_boxes_preprocessing(self, imgs, boxes): """ This function performs preprocessing for the input images and corresponding boxes for one clip. Args: imgs (tensor): the images. boxes (ndarray): the boxes for the current clip. Returns: imgs (tensor): list of preprocessed images. boxes (ndarray): preprocessed boxes. """ # Image [0, 255] -> [0, 1]. imgs = imgs.float() imgs = imgs / 255.0 height, width = imgs.shape[2], imgs.shape[3] # The format of boxes is [x1, y1, x2, y2]. The input boxes are in the # range of [0, 1]. boxes[:, [0, 2]] *= width boxes[:, [1, 3]] *= height boxes = transform.clip_boxes_to_image(boxes, height, width) if self._split == "train": # Train split imgs, boxes = transform.random_short_side_scale_jitter( imgs, min_size=self._jitter_min_scale, max_size=self._jitter_max_scale, boxes=boxes, ) imgs, boxes = transform.random_crop( imgs, self._crop_size, boxes=boxes ) # Random flip. imgs, boxes = transform.horizontal_flip(0.5, imgs, boxes=boxes) elif self._split == "val": # Val split # Resize short side to crop_size. Non-local and STRG uses 256. imgs, boxes = transform.random_short_side_scale_jitter( imgs, min_size=self._crop_size, max_size=self._crop_size, boxes=boxes, ) # Apply center crop for val split imgs, boxes = transform.uniform_crop( imgs, size=self._crop_size, spatial_idx=1, boxes=boxes ) if self._test_force_flip: imgs, boxes = transform.horizontal_flip(1, imgs, boxes=boxes) elif self._split == "test": # Test split # Resize short side to crop_size. Non-local and STRG uses 256. imgs, boxes = transform.random_short_side_scale_jitter( imgs, min_size=self._crop_size, max_size=self._crop_size, boxes=boxes, ) if self._test_force_flip: imgs, boxes = transform.horizontal_flip(1, imgs, boxes=boxes) else: raise NotImplementedError( "{} split not supported yet!".format(self._split) ) # Do color augmentation (after divided by 255.0). if self._split == "train" and self._use_color_augmentation: if not self._pca_jitter_only: imgs = transform.color_jitter( imgs, img_brightness=0.4, img_contrast=0.4, img_saturation=0.4, ) imgs = transform.lighting_jitter( imgs, alphastd=0.1, eigval=np.array(self._pca_eigval).astype(np.float32), eigvec=np.array(self._pca_eigvec).astype(np.float32), ) # Normalize images by mean and std. imgs = transform.color_normalization( imgs, np.array(self._data_mean, dtype=np.float32), np.array(self._data_std, dtype=np.float32), ) if not self._use_bgr: # Convert image format from BGR to RGB. # Note that Kinetics pre-training uses RGB! imgs = imgs[:, [2, 1, 0], ...] boxes = transform.clip_boxes_to_image( boxes, self._crop_size, self._crop_size ) return imgs, boxes def __getitem__(self, idx): """ Generate corresponding clips, boxes, labels and metadata for given idx. Args: idx (int): the video index provided by the pytorch sampler. Returns: frames (tensor): the frames of sampled from the video. The dimension is `channel` x `num frames` x `height` x `width`. label (ndarray): the label for correspond boxes for the current video. idx (int): the video index provided by the pytorch sampler. extra_data (dict): a dict containing extra data fields, like "boxes", "ori_boxes" and "metadata". """ video_idx, sec_idx, sec, center_idx = self._keyframe_indices[idx] # Get the frame idxs for current clip. seq = utils.get_sequence( center_idx, self._seq_len // 2, self._sample_rate, num_frames=len(self._image_paths[video_idx]), ) clip_label_list = self._keyframe_boxes_and_labels[video_idx][sec_idx] assert len(clip_label_list) > 0 # Get boxes and labels for current clip. boxes = [] labels = [] for box_labels in clip_label_list: boxes.append(box_labels[0]) labels.append(box_labels[1]) boxes = np.array(boxes) # Score is not used. boxes = boxes[:, :4].copy() ori_boxes = boxes.copy() # Load images of current clip. image_paths = [self._image_paths[video_idx][frame] for frame in seq] imgs = utils.retry_load_images( image_paths, backend=self.cfg.AVA.IMG_PROC_BACKEND ) if self.cfg.AVA.IMG_PROC_BACKEND == "pytorch": # T H W C -> T C H W. imgs = imgs.permute(0, 3, 1, 2) # Preprocess images and boxes. imgs, boxes = self._images_and_boxes_preprocessing( imgs, boxes=boxes ) # T C H W -> C T H W. imgs = imgs.permute(1, 0, 2, 3) else: # Preprocess images and boxes imgs, boxes = self._images_and_boxes_preprocessing_cv2( imgs, boxes=boxes ) # Construct label arrays. label_arrs = np.zeros((len(labels), self._num_classes), dtype=np.int32) for i, box_labels in enumerate(labels): # AVA label index starts from 1. for label in box_labels: if label == -1: continue assert label >= 1 and label <= 80 label_arrs[i][label - 1] = 1 imgs = utils.pack_pathway_output(self.cfg, imgs) metadata = [[video_idx, sec]] * len(boxes) extra_data = { "boxes": boxes, "ori_boxes": ori_boxes, "metadata": metadata, } return imgs, label_arrs, idx, extra_data
14,963
33.881119
82
py
STTS
STTS-main/MViT/slowfast/datasets/ptv_datasets.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import functools import os from typing import Dict import torch from torch.utils.data import ( DistributedSampler, RandomSampler, SequentialSampler, ) from torchvision.transforms import Compose, Lambda from torchvision.transforms._transforms_video import ( NormalizeVideo, RandomCropVideo, RandomHorizontalFlipVideo, ) import slowfast.utils.logging as logging from pytorchvideo.data import ( Charades, LabeledVideoDataset, SSv2, make_clip_sampler, ) from pytorchvideo.data.labeled_video_paths import LabeledVideoPaths from pytorchvideo.transforms import ( ApplyTransformToKey, RandomShortSideScale, ShortSideScale, UniformCropVideo, UniformTemporalSubsample, ) from . import utils as utils from .build import DATASET_REGISTRY logger = logging.get_logger(__name__) class PTVDatasetWrapper(torch.utils.data.IterableDataset): """ Wrapper for PyTorchVideo datasets. """ def __init__(self, num_videos, clips_per_video, crops_per_clip, dataset): """ Construct the dataset. Args: num_vidoes (int): number of videos in the dataset. clips_per_video (int): number of clips per video in the dataset. dataset (torch.utils.data.IterableDataset): a PyTorchVideo dataset. """ self._clips_per_video = clips_per_video self._crops_per_clip = crops_per_clip self._num_videos = num_videos self.dataset = dataset def __next__(self): """ Retrieves the next clip from the dataset. """ return self.dataset.__next__() @property def sampler(self): """ Returns: (torch.utils.data.Sampler): video sampler for the dataset. """ return self.dataset.video_sampler def __len__(self): """ Returns: (int): the number of clips per replica in the IterableDataset. """ return len(self.sampler) * self._clips_per_video * self._crops_per_clip @property def num_videos(self): """ Returns: (int): the number of clips in total in the dataset. """ return self._num_videos * self._clips_per_video * self._crops_per_clip def __iter__(self): return self class PackPathway(torch.nn.Module): """ Transform for converting video frames as a list of tensors. Each tensor corresponding to a unique pathway. """ def __init__(self, cfg): super().__init__() self.cfg = cfg def forward(self, x: torch.Tensor): return utils.pack_pathway_output(self.cfg, x) class DictToTuple(torch.nn.Module): """ Transform for converting output from dict to a tuple following PySlowFast dataset output format. """ def __init__(self, num_clips, num_crops): super().__init__() self._num_clips = num_clips self._num_crops = num_crops def forward(self, x: Dict[str, torch.Tensor]): index = ( x["video_index"] * self._num_clips * self._num_crops + x["clip_index"] * self._num_crops + x["aug_index"] ) return x["video"], x["label"], index, {} def div255(x): """ Scale clip frames from [0, 255] to [0, 1]. Args: x (Tensor): A tensor of the clip's RGB frames with shape: (channel, time, height, width). Returns: x (Tensor): Scaled tensor by divide 255. """ return x / 255.0 @DATASET_REGISTRY.register() def Ptvkinetics(cfg, mode): """ Construct the Kinetics video loader with a given csv file. The format of the csv file is: ``` path_to_video_1 label_1 path_to_video_2 label_2 ... path_to_video_N label_N ``` For `train` and `val` mode, a single clip is randomly sampled from every video with random cropping, scaling, and flipping. For `test` mode, multiple clips are uniformaly sampled from every video with center cropping. Args: cfg (CfgNode): configs. mode (string): Options includes `train`, `val`, or `test` mode. For the train and val mode, the data loader will take data from the train or val set, and sample one clip per video. For the test mode, the data loader will take data from test set, and sample multiple clips per video. """ # Only support train, val, and test mode. assert mode in [ "train", "val", "test", ], "Split '{}' not supported".format(mode) logger.info("Constructing Ptvkinetics {}...".format(mode)) clip_duration = ( cfg.DATA.NUM_FRAMES * cfg.DATA.SAMPLING_RATE / cfg.DATA.TARGET_FPS ) path_to_file = os.path.join( cfg.DATA.PATH_TO_DATA_DIR, "{}.csv".format(mode) ) labeled_video_paths = LabeledVideoPaths.from_path(path_to_file) num_videos = len(labeled_video_paths) labeled_video_paths.path_prefix = cfg.DATA.PATH_PREFIX logger.info( "Constructing kinetics dataloader (size: {}) from {}".format( num_videos, path_to_file ) ) if mode in ["train", "val"]: num_clips = 1 num_crops = 1 transform = Compose( [ ApplyTransformToKey( key="video", transform=Compose( [ UniformTemporalSubsample(cfg.DATA.NUM_FRAMES), Lambda(div255), NormalizeVideo(cfg.DATA.MEAN, cfg.DATA.STD), RandomShortSideScale( min_size=cfg.DATA.TRAIN_JITTER_SCALES[0], max_size=cfg.DATA.TRAIN_JITTER_SCALES[1], ), RandomCropVideo(cfg.DATA.TRAIN_CROP_SIZE), ] + ( [RandomHorizontalFlipVideo(p=0.5)] if cfg.DATA.RANDOM_FLIP else [] ) + [PackPathway(cfg)] ), ), DictToTuple(num_clips, num_crops), ] ) clip_sampler = make_clip_sampler("random", clip_duration) if cfg.NUM_GPUS > 1: video_sampler = DistributedSampler else: video_sampler = ( RandomSampler if mode == "train" else SequentialSampler ) else: num_clips = cfg.TEST.NUM_ENSEMBLE_VIEWS num_crops = cfg.TEST.NUM_SPATIAL_CROPS transform = Compose( [ ApplyTransformToKey( key="video", transform=Compose( [ UniformTemporalSubsample(cfg.DATA.NUM_FRAMES), Lambda(div255), NormalizeVideo(cfg.DATA.MEAN, cfg.DATA.STD), ShortSideScale( size=cfg.DATA.TRAIN_JITTER_SCALES[0] ), ] ), ), UniformCropVideo(size=cfg.DATA.TEST_CROP_SIZE), ApplyTransformToKey(key="video", transform=PackPathway(cfg)), DictToTuple(num_clips, num_crops), ] ) clip_sampler = make_clip_sampler( "constant_clips_per_video", clip_duration, num_clips, num_crops, ) video_sampler = ( DistributedSampler if cfg.NUM_GPUS > 1 else SequentialSampler ) return PTVDatasetWrapper( num_videos=num_videos, clips_per_video=num_clips, crops_per_clip=num_crops, dataset=LabeledVideoDataset( labeled_video_paths=labeled_video_paths, clip_sampler=clip_sampler, video_sampler=video_sampler, transform=transform, decode_audio=False, ), ) def process_charades_label(x, mode, num_classes): """ Process the video label for Charades dataset. Use video-level label for training mode, otherwise use clip-level label. Then convert the label into a binary vector. Args: x (dict): a video clip including label index. mode (string): Options includes `train`, `val`, or `test` mode. num_classes (int): Number of classes in the dataset. Returns: x (dict): video clip with updated label information. """ label = ( utils.aggregate_labels(x["label"]) if mode == "train" else x["video_label"] ) x["label"] = torch.as_tensor(utils.as_binary_vector(label, num_classes)) return x def rgb2bgr(x): """ Convert clip frames from RGB mode to BRG mode. Args: x (Tensor): A tensor of the clip's RGB frames with shape: (channel, time, height, width). Returns: x (Tensor): Converted tensor """ return x[[2, 1, 0], ...] @DATASET_REGISTRY.register() def Ptvcharades(cfg, mode): """ Construct PyTorchVideo Charades video loader. Load Charades data (frame paths, labels, etc. ) to Charades Dataset object. The dataset could be downloaded from Chrades official website (https://allenai.org/plato/charades/). Please see datasets/DATASET.md for more information about the data format. For `train` and `val` mode, a single clip is randomly sampled from every video with random cropping, scaling, and flipping. For `test` mode, multiple clips are uniformaly sampled from every video with center cropping. Args: cfg (CfgNode): configs. mode (string): Options includes `train`, `val`, or `test` mode. For the train and val mode, the data loader will take data from the train or val set, and sample one clip per video. For the test mode, the data loader will take data from test set, and sample multiple clips per video. """ # Only support train, val, and test mode. assert mode in [ "train", "val", "test", ], "Split '{}' not supported".format(mode) logger.info("Constructing Ptvcharades {}...".format(mode)) clip_duration = ( (cfg.DATA.NUM_FRAMES - 1) * cfg.DATA.SAMPLING_RATE + 1 ) / cfg.DATA.TARGET_FPS if mode in ["train", "val"]: num_clips = 1 num_crops = 1 transform = Compose( [ ApplyTransformToKey( key="video", transform=Compose( [ Lambda(div255), NormalizeVideo(cfg.DATA.MEAN, cfg.DATA.STD), RandomShortSideScale( min_size=cfg.DATA.TRAIN_JITTER_SCALES[0], max_size=cfg.DATA.TRAIN_JITTER_SCALES[1], ), RandomCropVideo(cfg.DATA.TRAIN_CROP_SIZE), Lambda(rgb2bgr), ] + ( [RandomHorizontalFlipVideo(p=0.5)] if cfg.DATA.RANDOM_FLIP else [] ) + [PackPathway(cfg)] ), ), Lambda( functools.partial( process_charades_label, mode=mode, num_classes=cfg.MODEL.NUM_CLASSES, ) ), DictToTuple(num_clips, num_crops), ] ) clip_sampler = make_clip_sampler("random", clip_duration) if cfg.NUM_GPUS > 1: video_sampler = DistributedSampler else: video_sampler = ( RandomSampler if mode == "train" else SequentialSampler ) else: num_clips = cfg.TEST.NUM_ENSEMBLE_VIEWS num_crops = cfg.TEST.NUM_SPATIAL_CROPS transform = Compose( [ ApplyTransformToKey( key="video", transform=Compose( [ Lambda(div255), NormalizeVideo(cfg.DATA.MEAN, cfg.DATA.STD), ShortSideScale(size=cfg.DATA.TEST_CROP_SIZE), ] ), ), UniformCropVideo(size=cfg.DATA.TEST_CROP_SIZE), Lambda( functools.partial( process_charades_label, mode=mode, num_classes=cfg.MODEL.NUM_CLASSES, ) ), ApplyTransformToKey( key="video", transform=Compose( [Lambda(rgb2bgr), PackPathway(cfg)], ), ), DictToTuple(num_clips, num_crops), ] ) clip_sampler = make_clip_sampler( "constant_clips_per_video", clip_duration, num_clips, num_crops, ) video_sampler = ( DistributedSampler if cfg.NUM_GPUS > 1 else SequentialSampler ) data_path = os.path.join(cfg.DATA.PATH_TO_DATA_DIR, "{}.csv".format(mode)) dataset = Charades( data_path=data_path, clip_sampler=clip_sampler, video_sampler=video_sampler, transform=transform, video_path_prefix=cfg.DATA.PATH_PREFIX, frames_per_clip=cfg.DATA.NUM_FRAMES, ) logger.info( "Constructing charades dataloader (size: {}) from {}".format( len(dataset._path_to_videos), data_path ) ) return PTVDatasetWrapper( num_videos=len(dataset._path_to_videos), clips_per_video=num_clips, crops_per_clip=num_crops, dataset=dataset, ) @DATASET_REGISTRY.register() def Ptvssv2(cfg, mode): """ Construct PyTorchVideo Something-Something v2 SSv2 video loader. Load SSv2 data (frame paths, labels, etc. ) to SSv2 Dataset object. The dataset could be downloaded from Chrades official website (https://20bn.com/datasets/something-something). Please see datasets/DATASET.md for more information about the data format. For training and validation, a single clip is randomly sampled from every video with random cropping and scaling. For testing, multiple clips are uniformaly sampled from every video with uniform cropping. For uniform cropping, we take the left, center, and right crop if the width is larger than height, or take top, center, and bottom crop if the height is larger than the width. Args: cfg (CfgNode): configs. mode (string): Options includes `train`, `val`, or `test` mode. """ # Only support train, val, and test mode. assert mode in [ "train", "val", "test", ], "Split '{}' not supported".format(mode) logger.info("Constructing Ptvcharades {}...".format(mode)) if mode in ["train", "val"]: num_clips = 1 num_crops = 1 transform = Compose( [ ApplyTransformToKey( key="video", transform=Compose( [ Lambda(div255), NormalizeVideo(cfg.DATA.MEAN, cfg.DATA.STD), RandomShortSideScale( min_size=cfg.DATA.TRAIN_JITTER_SCALES[0], max_size=cfg.DATA.TRAIN_JITTER_SCALES[1], ), RandomCropVideo(cfg.DATA.TRAIN_CROP_SIZE), Lambda(rgb2bgr), ] + ( [RandomHorizontalFlipVideo(p=0.5)] if cfg.DATA.RANDOM_FLIP else [] ) + [PackPathway(cfg)] ), ), DictToTuple(num_clips, num_crops), ] ) clip_sampler = make_clip_sampler( "constant_clips_per_video", 1, # Put arbitrary duration as ssv2 always needs full video clip. num_clips, num_crops, ) if cfg.NUM_GPUS > 1: video_sampler = DistributedSampler else: video_sampler = ( RandomSampler if mode == "train" else SequentialSampler ) else: assert cfg.TEST.NUM_ENSEMBLE_VIEWS == 1 num_clips = cfg.TEST.NUM_ENSEMBLE_VIEWS num_crops = cfg.TEST.NUM_SPATIAL_CROPS transform = Compose( [ ApplyTransformToKey( key="video", transform=Compose( [ Lambda(div255), NormalizeVideo(cfg.DATA.MEAN, cfg.DATA.STD), ShortSideScale(size=cfg.DATA.TEST_CROP_SIZE), ] ), ), UniformCropVideo(size=cfg.DATA.TEST_CROP_SIZE), ApplyTransformToKey( key="video", transform=Compose( [Lambda(rgb2bgr), PackPathway(cfg)], ), ), DictToTuple(num_clips, num_crops), ] ) clip_sampler = make_clip_sampler( "constant_clips_per_video", 1, # Put arbitrary duration as ssv2 always needs full video clip. num_clips, num_crops, ) video_sampler = ( DistributedSampler if cfg.NUM_GPUS > 1 else SequentialSampler ) label_name_file = os.path.join( cfg.DATA.PATH_TO_DATA_DIR, "something-something-v2-labels.json" ) video_label_file = os.path.join( cfg.DATA.PATH_TO_DATA_DIR, "something-something-v2-{}.json".format( "train" if mode == "train" else "validation" ), ) data_path = os.path.join( cfg.DATA.PATH_TO_DATA_DIR, "{}.csv".format("train" if mode == "train" else "val"), ) dataset = SSv2( label_name_file=label_name_file, video_label_file=video_label_file, video_path_label_file=data_path, clip_sampler=clip_sampler, video_sampler=video_sampler, transform=transform, video_path_prefix=cfg.DATA.PATH_PREFIX, frames_per_clip=cfg.DATA.NUM_FRAMES, rand_sample_frames=mode == "train", ) logger.info( "Constructing ssv2 dataloader (size: {}) from {}".format( len(dataset._path_to_videos), data_path ) ) return PTVDatasetWrapper( num_videos=len(dataset._path_to_videos), clips_per_video=num_clips, crops_per_clip=num_crops, dataset=dataset, )
19,371
31.557983
84
py
STTS
STTS-main/MViT/slowfast/datasets/charades.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import os import random from itertools import chain as chain import torch import torch.utils.data import slowfast.utils.logging as logging from slowfast.utils.env import pathmgr from . import utils as utils from .build import DATASET_REGISTRY logger = logging.get_logger(__name__) @DATASET_REGISTRY.register() class Charades(torch.utils.data.Dataset): """ Charades video loader. Construct the Charades video loader, then sample clips from the videos. For training and validation, a single clip is randomly sampled from every video with random cropping, scaling, and flipping. For testing, multiple clips are uniformaly sampled from every video with uniform cropping. For uniform cropping, we take the left, center, and right crop if the width is larger than height, or take top, center, and bottom crop if the height is larger than the width. """ def __init__(self, cfg, mode, num_retries=10): """ Load Charades data (frame paths, labels, etc. ) to a given Dataset object. The dataset could be downloaded from Chrades official website (https://allenai.org/plato/charades/). Please see datasets/DATASET.md for more information about the data format. Args: dataset (Dataset): a Dataset object to load Charades data to. mode (string): 'train', 'val', or 'test'. Args: cfg (CfgNode): configs. mode (string): Options includes `train`, `val`, or `test` mode. For the train and val mode, the data loader will take data from the train or val set, and sample one clip per video. For the test mode, the data loader will take data from test set, and sample multiple clips per video. num_retries (int): number of retries. """ # Only support train, val, and test mode. assert mode in [ "train", "val", "test", ], "Split '{}' not supported for Charades ".format(mode) self.mode = mode self.cfg = cfg self._video_meta = {} self._num_retries = num_retries # For training or validation mode, one single clip is sampled from every # video. For testing, NUM_ENSEMBLE_VIEWS clips are sampled from every # video. For every clip, NUM_SPATIAL_CROPS is cropped spatially from # the frames. if self.mode in ["train", "val"]: self._num_clips = 1 elif self.mode in ["test"]: self._num_clips = ( cfg.TEST.NUM_ENSEMBLE_VIEWS * cfg.TEST.NUM_SPATIAL_CROPS ) logger.info("Constructing Charades {}...".format(mode)) self._construct_loader() def _construct_loader(self): """ Construct the video loader. """ path_to_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "{}.csv".format("train" if self.mode == "train" else "val"), ) assert pathmgr.exists(path_to_file), "{} dir not found".format( path_to_file ) (self._path_to_videos, self._labels) = utils.load_image_lists( path_to_file, self.cfg.DATA.PATH_PREFIX, return_list=True ) if self.mode != "train": # Form video-level labels from frame level annotations. self._labels = utils.convert_to_video_level_labels(self._labels) self._path_to_videos = list( chain.from_iterable( [[x] * self._num_clips for x in self._path_to_videos] ) ) self._labels = list( chain.from_iterable([[x] * self._num_clips for x in self._labels]) ) self._spatial_temporal_idx = list( chain.from_iterable( [range(self._num_clips) for _ in range(len(self._labels))] ) ) logger.info( "Charades dataloader constructed (size: {}) from {}".format( len(self._path_to_videos), path_to_file ) ) def get_seq_frames(self, index): """ Given the video index, return the list of indexs of sampled frames. Args: index (int): the video index. Returns: seq (list): the indexes of sampled frames from the video. """ temporal_sample_index = ( -1 if self.mode in ["train", "val"] else self._spatial_temporal_idx[index] // self.cfg.TEST.NUM_SPATIAL_CROPS ) num_frames = self.cfg.DATA.NUM_FRAMES sampling_rate = utils.get_random_sampling_rate( self.cfg.MULTIGRID.LONG_CYCLE_SAMPLING_RATE, self.cfg.DATA.SAMPLING_RATE, ) video_length = len(self._path_to_videos[index]) assert video_length == len(self._labels[index]) clip_length = (num_frames - 1) * sampling_rate + 1 if temporal_sample_index == -1: if clip_length > video_length: start = random.randint(video_length - clip_length, 0) else: start = random.randint(0, video_length - clip_length) else: gap = float(max(video_length - clip_length, 0)) / ( self.cfg.TEST.NUM_ENSEMBLE_VIEWS - 1 ) start = int(round(gap * temporal_sample_index)) seq = [ max(min(start + i * sampling_rate, video_length - 1), 0) for i in range(num_frames) ] return seq def __getitem__(self, index): """ Given the video index, return the list of frames, label, and video index if the video frames can be fetched. Args: index (int): the video index provided by the pytorch sampler. Returns: frames (tensor): the frames of sampled from the video. The dimension is `channel` x `num frames` x `height` x `width`. label (int): the label of the current video. index (int): the index of the video. """ short_cycle_idx = None # When short cycle is used, input index is a tupple. if isinstance(index, tuple): index, short_cycle_idx = index if self.mode in ["train", "val"]: # -1 indicates random sampling. spatial_sample_index = -1 min_scale = self.cfg.DATA.TRAIN_JITTER_SCALES[0] max_scale = self.cfg.DATA.TRAIN_JITTER_SCALES[1] crop_size = self.cfg.DATA.TRAIN_CROP_SIZE if short_cycle_idx in [0, 1]: crop_size = int( round( self.cfg.MULTIGRID.SHORT_CYCLE_FACTORS[short_cycle_idx] * self.cfg.MULTIGRID.DEFAULT_S ) ) if self.cfg.MULTIGRID.DEFAULT_S > 0: # Decreasing the scale is equivalent to using a larger "span" # in a sampling grid. min_scale = int( round( float(min_scale) * crop_size / self.cfg.MULTIGRID.DEFAULT_S ) ) elif self.mode in ["test"]: # spatial_sample_index is in [0, 1, 2]. Corresponding to left, # center, or right if width is larger than height, and top, middle, # or bottom if height is larger than width. spatial_sample_index = ( self._spatial_temporal_idx[index] % self.cfg.TEST.NUM_SPATIAL_CROPS ) min_scale, max_scale, crop_size = [self.cfg.DATA.TEST_CROP_SIZE] * 3 # The testing is deterministic and no jitter should be performed. # min_scale, max_scale, and crop_size are expect to be the same. assert len({min_scale, max_scale, crop_size}) == 1 else: raise NotImplementedError( "Does not support {} mode".format(self.mode) ) seq = self.get_seq_frames(index) frames = torch.as_tensor( utils.retry_load_images( [self._path_to_videos[index][frame] for frame in seq], self._num_retries, ) ) label = utils.aggregate_labels( [self._labels[index][i] for i in range(seq[0], seq[-1] + 1)] ) label = torch.as_tensor( utils.as_binary_vector(label, self.cfg.MODEL.NUM_CLASSES) ) # Perform color normalization. frames = utils.tensor_normalize( frames, self.cfg.DATA.MEAN, self.cfg.DATA.STD ) # T H W C -> C T H W. frames = frames.permute(3, 0, 1, 2) # Perform data augmentation. frames = utils.spatial_sampling( frames, spatial_idx=spatial_sample_index, min_scale=min_scale, max_scale=max_scale, crop_size=crop_size, random_horizontal_flip=self.cfg.DATA.RANDOM_FLIP, inverse_uniform_sampling=self.cfg.DATA.INV_UNIFORM_SAMPLE, ) frames = utils.pack_pathway_output(self.cfg, frames) return frames, label, index, {} def __len__(self): """ Returns: (int): the number of videos in the dataset. """ return self.num_videos @property def num_videos(self): """ Returns: (int): the number of videos in the dataset. """ return len(self._path_to_videos)
9,677
36.366795
82
py
STTS
STTS-main/MViT/slowfast/datasets/multigrid_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """Helper functions for multigrid training.""" import numpy as np import torch from torch.utils.data.sampler import Sampler TORCH_MAJOR = int(torch.__version__.split(".")[0]) TORCH_MINOR = int(torch.__version__.split(".")[1]) if TORCH_MAJOR >= 1 and TORCH_MINOR >= 8: _int_classes = int else: from torch._six import int_classes as _int_classes class ShortCycleBatchSampler(Sampler): """ Extend Sampler to support "short cycle" sampling. See paper "A Multigrid Method for Efficiently Training Video Models", Wu et al., 2019 (https://arxiv.org/abs/1912.00998) for details. """ def __init__(self, sampler, batch_size, drop_last, cfg): if not isinstance(sampler, Sampler): raise ValueError( "sampler should be an instance of " "torch.utils.data.Sampler, but got sampler={}".format(sampler) ) if ( not isinstance(batch_size, _int_classes) or isinstance(batch_size, bool) or batch_size <= 0 ): raise ValueError( "batch_size should be a positive integer value, " "but got batch_size={}".format(batch_size) ) if not isinstance(drop_last, bool): raise ValueError( "drop_last should be a boolean value, but got " "drop_last={}".format(drop_last) ) self.sampler = sampler self.drop_last = drop_last bs_factor = [ int( round( ( float(cfg.DATA.TRAIN_CROP_SIZE) / (s * cfg.MULTIGRID.DEFAULT_S) ) ** 2 ) ) for s in cfg.MULTIGRID.SHORT_CYCLE_FACTORS ] self.batch_sizes = [ batch_size * bs_factor[0], batch_size * bs_factor[1], batch_size, ] def __iter__(self): counter = 0 batch_size = self.batch_sizes[0] batch = [] for idx in self.sampler: batch.append((idx, counter % 3)) if len(batch) == batch_size: yield batch counter += 1 batch_size = self.batch_sizes[counter % 3] batch = [] if len(batch) > 0 and not self.drop_last: yield batch def __len__(self): avg_batch_size = sum(self.batch_sizes) / 3.0 if self.drop_last: return int(np.floor(len(self.sampler) / avg_batch_size)) else: return int(np.ceil(len(self.sampler) / avg_batch_size))
2,753
30.295455
78
py
STTS
STTS-main/MViT/slowfast/datasets/decoder.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import math import numpy as np import random import torch import torchvision.io as io def temporal_sampling(frames, start_idx, end_idx, num_samples): """ Given the start and end frame index, sample num_samples frames between the start and end with equal interval. Args: frames (tensor): a tensor of video frames, dimension is `num video frames` x `channel` x `height` x `width`. start_idx (int): the index of the start frame. end_idx (int): the index of the end frame. num_samples (int): number of frames to sample. Returns: frames (tersor): a tensor of temporal sampled video frames, dimension is `num clip frames` x `channel` x `height` x `width`. """ index = torch.linspace(start_idx, end_idx, num_samples) index = torch.clamp(index, 0, frames.shape[0] - 1).long() frames = torch.index_select(frames, 0, index) return frames def get_start_end_idx( video_size, clip_size, clip_idx, num_clips, use_offset=False ): """ Sample a clip of size clip_size from a video of size video_size and return the indices of the first and last frame of the clip. If clip_idx is -1, the clip is randomly sampled, otherwise uniformly split the video to num_clips clips, and select the start and end index of clip_idx-th video clip. Args: video_size (int): number of overall frames. clip_size (int): size of the clip to sample from the frames. clip_idx (int): if clip_idx is -1, perform random jitter sampling. If clip_idx is larger than -1, uniformly split the video to num_clips clips, and select the start and end index of the clip_idx-th video clip. num_clips (int): overall number of clips to uniformly sample from the given video for testing. Returns: start_idx (int): the start frame index. end_idx (int): the end frame index. """ delta = max(video_size - clip_size, 0) if clip_idx == -1: # Random temporal sampling. start_idx = random.uniform(0, delta) else: if use_offset: if num_clips == 1: # Take the center clip if num_clips is 1. start_idx = math.floor(delta / 2) else: # Uniformly sample the clip with the given index. start_idx = clip_idx * math.floor(delta / (num_clips - 1)) else: # Uniformly sample the clip with the given index. start_idx = delta * clip_idx / num_clips end_idx = start_idx + clip_size - 1 return start_idx, end_idx def pyav_decode_stream( container, start_pts, end_pts, stream, stream_name, buffer_size=0 ): """ Decode the video with PyAV decoder. Args: container (container): PyAV container. start_pts (int): the starting Presentation TimeStamp to fetch the video frames. end_pts (int): the ending Presentation TimeStamp of the decoded frames. stream (stream): PyAV stream. stream_name (dict): a dictionary of streams. For example, {"video": 0} means video stream at stream index 0. buffer_size (int): number of additional frames to decode beyond end_pts. Returns: result (list): list of frames decoded. max_pts (int): max Presentation TimeStamp of the video sequence. """ # Seeking in the stream is imprecise. Thus, seek to an ealier PTS by a # margin pts. margin = 1024 seek_offset = max(start_pts - margin, 0) container.seek(seek_offset, any_frame=False, backward=True, stream=stream) frames = {} buffer_count = 0 max_pts = 0 for frame in container.decode(**stream_name): max_pts = max(max_pts, frame.pts) if frame.pts < start_pts: continue if frame.pts <= end_pts: frames[frame.pts] = frame else: buffer_count += 1 frames[frame.pts] = frame if buffer_count >= buffer_size: break result = [frames[pts] for pts in sorted(frames)] return result, max_pts def torchvision_decode( video_handle, sampling_rate, num_frames, clip_idx, video_meta, num_clips=10, target_fps=30, modalities=("visual",), max_spatial_scale=0, use_offset=False, ): """ If video_meta is not empty, perform temporal selective decoding to sample a clip from the video with TorchVision decoder. If video_meta is empty, decode the entire video and update the video_meta. Args: video_handle (bytes): raw bytes of the video file. sampling_rate (int): frame sampling rate (interval between two sampled frames). num_frames (int): number of frames to sample. clip_idx (int): if clip_idx is -1, perform random temporal sampling. If clip_idx is larger than -1, uniformly split the video to num_clips clips, and select the clip_idx-th video clip. video_meta (dict): a dict contains VideoMetaData. Details can be found at `pytorch/vision/torchvision/io/_video_opt.py`. num_clips (int): overall number of clips to uniformly sample from the given video. target_fps (int): the input video may has different fps, convert it to the target video fps. modalities (tuple): tuple of modalities to decode. Currently only support `visual`, planning to support `acoustic` soon. max_spatial_scale (int): the maximal resolution of the spatial shorter edge size during decoding. Returns: frames (tensor): decoded frames from the video. fps (float): the number of frames per second of the video. decode_all_video (bool): if True, the entire video was decoded. """ # Convert the bytes to a tensor. video_tensor = torch.from_numpy(np.frombuffer(video_handle, dtype=np.uint8)) decode_all_video = True video_start_pts, video_end_pts = 0, -1 # The video_meta is empty, fetch the meta data from the raw video. if len(video_meta) == 0: # Tracking the meta info for selective decoding in the future. meta = io._probe_video_from_memory(video_tensor) # Using the information from video_meta to perform selective decoding. video_meta["video_timebase"] = meta.video_timebase video_meta["video_numerator"] = meta.video_timebase.numerator video_meta["video_denominator"] = meta.video_timebase.denominator video_meta["has_video"] = meta.has_video video_meta["video_duration"] = meta.video_duration video_meta["video_fps"] = meta.video_fps video_meta["audio_timebas"] = meta.audio_timebase video_meta["audio_numerator"] = meta.audio_timebase.numerator video_meta["audio_denominator"] = meta.audio_timebase.denominator video_meta["has_audio"] = meta.has_audio video_meta["audio_duration"] = meta.audio_duration video_meta["audio_sample_rate"] = meta.audio_sample_rate fps = video_meta["video_fps"] if ( video_meta["has_video"] and video_meta["video_denominator"] > 0 and video_meta["video_duration"] > 0 ): # try selective decoding. decode_all_video = False clip_size = sampling_rate * num_frames / target_fps * fps start_idx, end_idx = get_start_end_idx( fps * video_meta["video_duration"], clip_size, clip_idx, num_clips, use_offset=use_offset, ) # Convert frame index to pts. pts_per_frame = video_meta["video_denominator"] / fps video_start_pts = int(start_idx * pts_per_frame) video_end_pts = int(end_idx * pts_per_frame) # Decode the raw video with the tv decoder. v_frames, _ = io._read_video_from_memory( video_tensor, seek_frame_margin=1.0, read_video_stream="visual" in modalities, video_width=0, video_height=0, video_min_dimension=max_spatial_scale, video_pts_range=(video_start_pts, video_end_pts), video_timebase_numerator=video_meta["video_numerator"], video_timebase_denominator=video_meta["video_denominator"], ) if v_frames.shape == torch.Size([0]): # failed selective decoding decode_all_video = True video_start_pts, video_end_pts = 0, -1 v_frames, _ = io._read_video_from_memory( video_tensor, seek_frame_margin=1.0, read_video_stream="visual" in modalities, video_width=0, video_height=0, video_min_dimension=max_spatial_scale, video_pts_range=(video_start_pts, video_end_pts), video_timebase_numerator=video_meta["video_numerator"], video_timebase_denominator=video_meta["video_denominator"], ) return v_frames, fps, decode_all_video def pyav_decode( container, sampling_rate, num_frames, clip_idx, num_clips=10, target_fps=30, use_offset=False, ): """ Convert the video from its original fps to the target_fps. If the video support selective decoding (contain decoding information in the video head), the perform temporal selective decoding and sample a clip from the video with the PyAV decoder. If the video does not support selective decoding, decode the entire video. Args: container (container): pyav container. sampling_rate (int): frame sampling rate (interval between two sampled frames. num_frames (int): number of frames to sample. clip_idx (int): if clip_idx is -1, perform random temporal sampling. If clip_idx is larger than -1, uniformly split the video to num_clips clips, and select the clip_idx-th video clip. num_clips (int): overall number of clips to uniformly sample from the given video. target_fps (int): the input video may has different fps, convert it to the target video fps before frame sampling. Returns: frames (tensor): decoded frames from the video. Return None if the no video stream was found. fps (float): the number of frames per second of the video. decode_all_video (bool): If True, the entire video was decoded. """ # Try to fetch the decoding information from the video head. Some of the # videos does not support fetching the decoding information, for that case # it will get None duration. fps = float(container.streams.video[0].average_rate) frames_length = container.streams.video[0].frames duration = container.streams.video[0].duration if duration is None: # If failed to fetch the decoding information, decode the entire video. decode_all_video = True video_start_pts, video_end_pts = 0, math.inf else: # Perform selective decoding. decode_all_video = False start_idx, end_idx = get_start_end_idx( frames_length, sampling_rate * num_frames / target_fps * fps, clip_idx, num_clips, use_offset=use_offset, ) timebase = duration / frames_length video_start_pts = int(start_idx * timebase) video_end_pts = int(end_idx * timebase) frames = None # If video stream was found, fetch video frames from the video. if container.streams.video: video_frames, max_pts = pyav_decode_stream( container, video_start_pts, video_end_pts, container.streams.video[0], {"video": 0}, ) container.close() frames = [frame.to_rgb().to_ndarray() for frame in video_frames] frames = torch.as_tensor(np.stack(frames)) return frames, fps, decode_all_video def decode( container, sampling_rate, num_frames, clip_idx=-1, num_clips=10, video_meta=None, target_fps=30, backend="pyav", max_spatial_scale=0, use_offset=False, ): """ Decode the video and perform temporal sampling. Args: container (container): pyav container. sampling_rate (int): frame sampling rate (interval between two sampled frames). num_frames (int): number of frames to sample. clip_idx (int): if clip_idx is -1, perform random temporal sampling. If clip_idx is larger than -1, uniformly split the video to num_clips clips, and select the clip_idx-th video clip. num_clips (int): overall number of clips to uniformly sample from the given video. video_meta (dict): a dict contains VideoMetaData. Details can be find at `pytorch/vision/torchvision/io/_video_opt.py`. target_fps (int): the input video may have different fps, convert it to the target video fps before frame sampling. backend (str): decoding backend includes `pyav` and `torchvision`. The default one is `pyav`. max_spatial_scale (int): keep the aspect ratio and resize the frame so that shorter edge size is max_spatial_scale. Only used in `torchvision` backend. Returns: frames (tensor): decoded frames from the video. """ # Currently support two decoders: 1) PyAV, and 2) TorchVision. assert clip_idx >= -1, "Not valied clip_idx {}".format(clip_idx) try: if backend == "pyav": frames, fps, decode_all_video = pyav_decode( container, sampling_rate, num_frames, clip_idx, num_clips, target_fps, use_offset=use_offset, ) elif backend == "torchvision": frames, fps, decode_all_video = torchvision_decode( container, sampling_rate, num_frames, clip_idx, video_meta, num_clips, target_fps, ("visual",), max_spatial_scale, use_offset=use_offset, ) else: raise NotImplementedError( "Unknown decoding backend {}".format(backend) ) except Exception as e: print("Failed to decode by {} with exception: {}".format(backend, e)) return None # Return None if the frames was not decoded successfully. if frames is None or frames.size(0) == 0: return None clip_sz = sampling_rate * num_frames / target_fps * fps start_idx, end_idx = get_start_end_idx( frames.shape[0], clip_sz, clip_idx if decode_all_video else 0, num_clips if decode_all_video else 1, use_offset=use_offset, ) # Perform temporal sampling from the decoded video. frames = temporal_sampling(frames, start_idx, end_idx, num_frames) return frames
15,165
37.787724
80
py
STTS
STTS-main/MViT/slowfast/datasets/ava_helper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import logging import os from collections import defaultdict from slowfast.utils.env import pathmgr logger = logging.getLogger(__name__) FPS = 30 AVA_VALID_FRAMES = range(902, 1799) def load_image_lists(cfg, is_train): """ Loading image paths from corresponding files. Args: cfg (CfgNode): config. is_train (bool): if it is training dataset or not. Returns: image_paths (list[list]): a list of items. Each item (also a list) corresponds to one video and contains the paths of images for this video. video_idx_to_name (list): a list which stores video names. """ list_filenames = [ os.path.join(cfg.AVA.FRAME_LIST_DIR, filename) for filename in ( cfg.AVA.TRAIN_LISTS if is_train else cfg.AVA.TEST_LISTS ) ] image_paths = defaultdict(list) video_name_to_idx = {} video_idx_to_name = [] for list_filename in list_filenames: with pathmgr.open(list_filename, "r") as f: f.readline() for line in f: row = line.split() # The format of each row should follow: # original_vido_id video_id frame_id path labels. assert len(row) == 5 video_name = row[0] if video_name not in video_name_to_idx: idx = len(video_name_to_idx) video_name_to_idx[video_name] = idx video_idx_to_name.append(video_name) data_key = video_name_to_idx[video_name] image_paths[data_key].append( os.path.join(cfg.AVA.FRAME_DIR, row[3]) ) image_paths = [image_paths[i] for i in range(len(image_paths))] logger.info( "Finished loading image paths from: %s" % ", ".join(list_filenames) ) return image_paths, video_idx_to_name def load_boxes_and_labels(cfg, mode): """ Loading boxes and labels from csv files. Args: cfg (CfgNode): config. mode (str): 'train', 'val', or 'test' mode. Returns: all_boxes (dict): a dict which maps from `video_name` and `frame_sec` to a list of `box`. Each `box` is a [`box_coord`, `box_labels`] where `box_coord` is the coordinates of box and 'box_labels` are the corresponding labels for the box. """ gt_lists = cfg.AVA.TRAIN_GT_BOX_LISTS if mode == "train" else [] pred_lists = ( cfg.AVA.TRAIN_PREDICT_BOX_LISTS if mode == "train" else cfg.AVA.TEST_PREDICT_BOX_LISTS ) ann_filenames = [ os.path.join(cfg.AVA.ANNOTATION_DIR, filename) for filename in gt_lists + pred_lists ] ann_is_gt_box = [True] * len(gt_lists) + [False] * len(pred_lists) detect_thresh = cfg.AVA.DETECTION_SCORE_THRESH # Only select frame_sec % 4 = 0 samples for validation if not # set FULL_TEST_ON_VAL. boxes_sample_rate = ( 4 if mode == "val" and not cfg.AVA.FULL_TEST_ON_VAL else 1 ) all_boxes, count, unique_box_count = parse_bboxes_file( ann_filenames=ann_filenames, ann_is_gt_box=ann_is_gt_box, detect_thresh=detect_thresh, boxes_sample_rate=boxes_sample_rate, ) logger.info( "Finished loading annotations from: %s" % ", ".join(ann_filenames) ) logger.info("Detection threshold: {}".format(detect_thresh)) logger.info("Number of unique boxes: %d" % unique_box_count) logger.info("Number of annotations: %d" % count) return all_boxes def get_keyframe_data(boxes_and_labels): """ Getting keyframe indices, boxes and labels in the dataset. Args: boxes_and_labels (list[dict]): a list which maps from video_idx to a dict. Each dict `frame_sec` to a list of boxes and corresponding labels. Returns: keyframe_indices (list): a list of indices of the keyframes. keyframe_boxes_and_labels (list[list[list]]): a list of list which maps from video_idx and sec_idx to a list of boxes and corresponding labels. """ def sec_to_frame(sec): """ Convert time index (in second) to frame index. 0: 900 30: 901 """ return (sec - 900) * FPS keyframe_indices = [] keyframe_boxes_and_labels = [] count = 0 for video_idx in range(len(boxes_and_labels)): sec_idx = 0 keyframe_boxes_and_labels.append([]) for sec in boxes_and_labels[video_idx].keys(): if sec not in AVA_VALID_FRAMES: continue if len(boxes_and_labels[video_idx][sec]) > 0: keyframe_indices.append( (video_idx, sec_idx, sec, sec_to_frame(sec)) ) keyframe_boxes_and_labels[video_idx].append( boxes_and_labels[video_idx][sec] ) sec_idx += 1 count += 1 logger.info("%d keyframes used." % count) return keyframe_indices, keyframe_boxes_and_labels def get_num_boxes_used(keyframe_indices, keyframe_boxes_and_labels): """ Get total number of used boxes. Args: keyframe_indices (list): a list of indices of the keyframes. keyframe_boxes_and_labels (list[list[list]]): a list of list which maps from video_idx and sec_idx to a list of boxes and corresponding labels. Returns: count (int): total number of used boxes. """ count = 0 for video_idx, sec_idx, _, _ in keyframe_indices: count += len(keyframe_boxes_and_labels[video_idx][sec_idx]) return count def parse_bboxes_file( ann_filenames, ann_is_gt_box, detect_thresh, boxes_sample_rate=1 ): """ Parse AVA bounding boxes files. Args: ann_filenames (list of str(s)): a list of AVA bounding boxes annotation files. ann_is_gt_box (list of bools): a list of boolean to indicate whether the corresponding ann_file is ground-truth. `ann_is_gt_box[i]` correspond to `ann_filenames[i]`. detect_thresh (float): threshold for accepting predicted boxes, range [0, 1]. boxes_sample_rate (int): sample rate for test bounding boxes. Get 1 every `boxes_sample_rate`. """ all_boxes = {} count = 0 unique_box_count = 0 for filename, is_gt_box in zip(ann_filenames, ann_is_gt_box): with pathmgr.open(filename, "r") as f: for line in f: row = line.strip().split(",") # When we use predicted boxes to train/eval, we need to # ignore the boxes whose scores are below the threshold. if not is_gt_box: score = float(row[7]) if score < detect_thresh: continue video_name, frame_sec = row[0], int(row[1]) if frame_sec % boxes_sample_rate != 0: continue # Box with format [x1, y1, x2, y2] with a range of [0, 1] as float. box_key = ",".join(row[2:6]) box = list(map(float, row[2:6])) label = -1 if row[6] == "" else int(row[6]) if video_name not in all_boxes: all_boxes[video_name] = {} for sec in AVA_VALID_FRAMES: all_boxes[video_name][sec] = {} if box_key not in all_boxes[video_name][frame_sec]: all_boxes[video_name][frame_sec][box_key] = [box, []] unique_box_count += 1 all_boxes[video_name][frame_sec][box_key][1].append(label) if label != -1: count += 1 for video_name in all_boxes.keys(): for frame_sec in all_boxes[video_name].keys(): # Save in format of a list of [box_i, box_i_labels]. all_boxes[video_name][frame_sec] = list( all_boxes[video_name][frame_sec].values() ) return all_boxes, count, unique_box_count
8,177
33.361345
102
py
STTS
STTS-main/MViT/slowfast/datasets/__init__.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. from .ava_dataset import Ava # noqa from .build import DATASET_REGISTRY, build_dataset # noqa from .charades import Charades # noqa from .imagenet import Imagenet # noqa from .kinetics import Kinetics # noqa from .ssv2 import Ssv2 # noqa try: from .ptv_datasets import Ptvcharades, Ptvkinetics, Ptvssv2 # noqa except Exception: print("Please update your PyTorchVideo to latest master")
498
32.266667
71
py
STTS
STTS-main/MViT/slowfast/datasets/ssv2.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import json import numpy as np import os import random from itertools import chain as chain import torch import torch.utils.data import slowfast.utils.logging as logging from slowfast.utils.env import pathmgr from . import utils as utils from .build import DATASET_REGISTRY logger = logging.get_logger(__name__) @DATASET_REGISTRY.register() class Ssv2(torch.utils.data.Dataset): """ Something-Something v2 (SSV2) video loader. Construct the SSV2 video loader, then sample clips from the videos. For training and validation, a single clip is randomly sampled from every video with random cropping, scaling, and flipping. For testing, multiple clips are uniformaly sampled from every video with uniform cropping. For uniform cropping, we take the left, center, and right crop if the width is larger than height, or take top, center, and bottom crop if the height is larger than the width. """ def __init__(self, cfg, mode, num_retries=10): """ Load Something-Something V2 data (frame paths, labels, etc. ) to a given Dataset object. The dataset could be downloaded from Something-Something official website (https://20bn.com/datasets/something-something). Please see datasets/DATASET.md for more information about the data format. Args: cfg (CfgNode): configs. mode (string): Options includes `train`, `val`, or `test` mode. For the train and val mode, the data loader will take data from the train or val set, and sample one clip per video. For the test mode, the data loader will take data from test set, and sample multiple clips per video. num_retries (int): number of retries for reading frames from disk. """ # Only support train, val, and test mode. assert mode in [ "train", "val", "test", ], "Split '{}' not supported for Something-Something V2".format(mode) self.mode = mode self.cfg = cfg self._video_meta = {} self._num_retries = num_retries # For training or validation mode, one single clip is sampled from every # video. For testing, NUM_ENSEMBLE_VIEWS clips are sampled from every # video. For every clip, NUM_SPATIAL_CROPS is cropped spatially from # the frames. if self.mode in ["train", "val"]: self._num_clips = 1 elif self.mode in ["test"]: self._num_clips = ( cfg.TEST.NUM_ENSEMBLE_VIEWS * cfg.TEST.NUM_SPATIAL_CROPS ) logger.info("Constructing Something-Something V2 {}...".format(mode)) self._construct_loader() def _construct_loader(self): """ Construct the video loader. """ # Loading label names. with pathmgr.open( os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "something-something-v2-labels.json", ), "r", ) as f: label_dict = json.load(f) # Loading labels. label_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "something-something-v2-{}.json".format( "train" if self.mode == "train" else "validation" ), ) with pathmgr.open(label_file, "r") as f: label_json = json.load(f) self._video_names = [] self._labels = [] for video in label_json: video_name = video["id"] template = video["template"] template = template.replace("[", "") template = template.replace("]", "") label = int(label_dict[template]) self._video_names.append(video_name) self._labels.append(label) path_to_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "{}.csv".format("train" if self.mode == "train" else "val"), ) assert pathmgr.exists(path_to_file), "{} dir not found".format( path_to_file ) self._path_to_videos, _ = utils.load_image_lists( path_to_file, self.cfg.DATA.PATH_PREFIX ) assert len(self._path_to_videos) == len(self._video_names), ( len(self._path_to_videos), len(self._video_names), ) # From dict to list. new_paths, new_labels = [], [] for index in range(len(self._video_names)): if self._video_names[index] in self._path_to_videos: new_paths.append(self._path_to_videos[self._video_names[index]]) new_labels.append(self._labels[index]) self._labels = new_labels self._path_to_videos = new_paths # Extend self when self._num_clips > 1 (during testing). self._path_to_videos = list( chain.from_iterable( [[x] * self._num_clips for x in self._path_to_videos] ) ) self._labels = list( chain.from_iterable([[x] * self._num_clips for x in self._labels]) ) self._spatial_temporal_idx = list( chain.from_iterable( [ range(self._num_clips) for _ in range(len(self._path_to_videos)) ] ) ) logger.info( "Something-Something V2 dataloader constructed " " (size: {}) from {}".format( len(self._path_to_videos), path_to_file ) ) def get_seq_frames(self, index): """ Given the video index, return the list of sampled frame indexes. Args: index (int): the video index. Returns: seq (list): the indexes of frames of sampled from the video. """ num_frames = self.cfg.DATA.NUM_FRAMES video_length = len(self._path_to_videos[index]) seg_size = float(video_length - 1) / num_frames seq = [] for i in range(num_frames): start = int(np.round(seg_size * i)) end = int(np.round(seg_size * (i + 1))) if self.mode == "train": seq.append(random.randint(start, end)) else: seq.append((start + end) // 2) return seq def __getitem__(self, index): """ Given the video index, return the list of frames, label, and video index if the video frames can be fetched. Args: index (int): the video index provided by the pytorch sampler. Returns: frames (tensor): the frames of sampled from the video. The dimension is `channel` x `num frames` x `height` x `width`. label (int): the label of the current video. index (int): the index of the video. """ short_cycle_idx = None # When short cycle is used, input index is a tupple. if isinstance(index, tuple): index, short_cycle_idx = index if self.mode in ["train", "val"]: # -1 indicates random sampling. spatial_sample_index = -1 min_scale = self.cfg.DATA.TRAIN_JITTER_SCALES[0] max_scale = self.cfg.DATA.TRAIN_JITTER_SCALES[1] crop_size = self.cfg.DATA.TRAIN_CROP_SIZE if short_cycle_idx in [0, 1]: crop_size = int( round( self.cfg.MULTIGRID.SHORT_CYCLE_FACTORS[short_cycle_idx] * self.cfg.MULTIGRID.DEFAULT_S ) ) if self.cfg.MULTIGRID.DEFAULT_S > 0: # Decreasing the scale is equivalent to using a larger "span" # in a sampling grid. min_scale = int( round( float(min_scale) * crop_size / self.cfg.MULTIGRID.DEFAULT_S ) ) elif self.mode in ["test"]: # spatial_sample_index is in [0, 1, 2]. Corresponding to left, # center, or right if width is larger than height, and top, middle, # or bottom if height is larger than width. spatial_sample_index = ( self._spatial_temporal_idx[index] % self.cfg.TEST.NUM_SPATIAL_CROPS ) min_scale, max_scale, crop_size = [self.cfg.DATA.TEST_CROP_SIZE] * 3 # The testing is deterministic and no jitter should be performed. # min_scale, max_scale, and crop_size are expect to be the same. assert len({min_scale, max_scale, crop_size}) == 1 else: raise NotImplementedError( "Does not support {} mode".format(self.mode) ) label = self._labels[index] seq = self.get_seq_frames(index) frames = torch.as_tensor( utils.retry_load_images( [self._path_to_videos[index][frame] for frame in seq], self._num_retries, ) ) # Perform color normalization. frames = utils.tensor_normalize( frames, self.cfg.DATA.MEAN, self.cfg.DATA.STD ) # T H W C -> C T H W. frames = frames.permute(3, 0, 1, 2) # Perform data augmentation. frames = utils.spatial_sampling( frames, spatial_idx=spatial_sample_index, min_scale=min_scale, max_scale=max_scale, crop_size=crop_size, random_horizontal_flip=self.cfg.DATA.RANDOM_FLIP, inverse_uniform_sampling=self.cfg.DATA.INV_UNIFORM_SAMPLE, ) frames = utils.pack_pathway_output(self.cfg, frames) return frames, label, index, {} def __len__(self): """ Returns: (int): the number of videos in the dataset. """ return self.num_videos @property def num_videos(self): """ Returns: (int): the number of videos in the dataset. """ return len(self._path_to_videos)
10,293
35.246479
82
py
STTS
STTS-main/MViT/slowfast/datasets/random_erasing.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. """ This implementation is based on https://github.com/rwightman/pytorch-image-models/blob/master/timm/data/random_erasing.py pulished under an Apache License 2.0. COMMENT FROM ORIGINAL: Originally inspired by impl at https://github.com/zhunzhong07/Random-Erasing, Apache 2.0 Copyright Zhun Zhong & Liang Zheng Hacked together by / Copyright 2020 Ross Wightman """ import math import random import torch def _get_pixels( per_pixel, rand_color, patch_size, dtype=torch.float32, device="cuda" ): # NOTE I've seen CUDA illegal memory access errors being caused by the normal_() # paths, flip the order so normal is run on CPU if this becomes a problem # Issue has been fixed in master https://github.com/pytorch/pytorch/issues/19508 if per_pixel: return torch.empty(patch_size, dtype=dtype, device=device).normal_() elif rand_color: return torch.empty( (patch_size[0], 1, 1), dtype=dtype, device=device ).normal_() else: return torch.zeros((patch_size[0], 1, 1), dtype=dtype, device=device) class RandomErasing: """Randomly selects a rectangle region in an image and erases its pixels. 'Random Erasing Data Augmentation' by Zhong et al. See https://arxiv.org/pdf/1708.04896.pdf This variant of RandomErasing is intended to be applied to either a batch or single image tensor after it has been normalized by dataset mean and std. Args: probability: Probability that the Random Erasing operation will be performed. min_area: Minimum percentage of erased area wrt input image area. max_area: Maximum percentage of erased area wrt input image area. min_aspect: Minimum aspect ratio of erased area. mode: pixel color mode, one of 'const', 'rand', or 'pixel' 'const' - erase block is constant color of 0 for all channels 'rand' - erase block is same per-channel random (normal) color 'pixel' - erase block is per-pixel random (normal) color max_count: maximum number of erasing blocks per image, area per box is scaled by count. per-image count is randomly chosen between 1 and this value. """ def __init__( self, probability=0.5, min_area=0.02, max_area=1 / 3, min_aspect=0.3, max_aspect=None, mode="const", min_count=1, max_count=None, num_splits=0, device="cuda", cube=True, ): self.probability = probability self.min_area = min_area self.max_area = max_area max_aspect = max_aspect or 1 / min_aspect self.log_aspect_ratio = (math.log(min_aspect), math.log(max_aspect)) self.min_count = min_count self.max_count = max_count or min_count self.num_splits = num_splits mode = mode.lower() self.rand_color = False self.per_pixel = False self.cube = cube if mode == "rand": self.rand_color = True # per block random normal elif mode == "pixel": self.per_pixel = True # per pixel random normal else: assert not mode or mode == "const" self.device = device def _erase(self, img, chan, img_h, img_w, dtype): if random.random() > self.probability: return area = img_h * img_w count = ( self.min_count if self.min_count == self.max_count else random.randint(self.min_count, self.max_count) ) for _ in range(count): for _ in range(10): target_area = ( random.uniform(self.min_area, self.max_area) * area / count ) aspect_ratio = math.exp(random.uniform(*self.log_aspect_ratio)) h = int(round(math.sqrt(target_area * aspect_ratio))) w = int(round(math.sqrt(target_area / aspect_ratio))) if w < img_w and h < img_h: top = random.randint(0, img_h - h) left = random.randint(0, img_w - w) img[:, top : top + h, left : left + w] = _get_pixels( self.per_pixel, self.rand_color, (chan, h, w), dtype=dtype, device=self.device, ) break def _erase_cube( self, img, batch_start, batch_size, chan, img_h, img_w, dtype, ): if random.random() > self.probability: return area = img_h * img_w count = ( self.min_count if self.min_count == self.max_count else random.randint(self.min_count, self.max_count) ) for _ in range(count): for _ in range(100): target_area = ( random.uniform(self.min_area, self.max_area) * area / count ) aspect_ratio = math.exp(random.uniform(*self.log_aspect_ratio)) h = int(round(math.sqrt(target_area * aspect_ratio))) w = int(round(math.sqrt(target_area / aspect_ratio))) if w < img_w and h < img_h: top = random.randint(0, img_h - h) left = random.randint(0, img_w - w) for i in range(batch_start, batch_size): img_instance = img[i] img_instance[ :, top : top + h, left : left + w ] = _get_pixels( self.per_pixel, self.rand_color, (chan, h, w), dtype=dtype, device=self.device, ) break def __call__(self, input): if len(input.size()) == 3: self._erase(input, *input.size(), input.dtype) else: batch_size, chan, img_h, img_w = input.size() # skip first slice of batch if num_splits is set (for clean portion of samples) batch_start = ( batch_size // self.num_splits if self.num_splits > 1 else 0 ) if self.cube: self._erase_cube( input, batch_start, batch_size, chan, img_h, img_w, input.dtype, ) else: for i in range(batch_start, batch_size): self._erase(input[i], chan, img_h, img_w, input.dtype) return input
6,887
37.055249
95
py
STTS
STTS-main/MViT/slowfast/datasets/cv2_transform.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import math import numpy as np import cv2 def clip_boxes_to_image(boxes, height, width): """ Clip the boxes with the height and width of the image size. Args: boxes (ndarray): bounding boxes to peform crop. The dimension is `num boxes` x 4. height (int): the height of the image. width (int): the width of the image. Returns: boxes (ndarray): cropped bounding boxes. """ boxes[:, [0, 2]] = np.minimum( width - 1.0, np.maximum(0.0, boxes[:, [0, 2]]) ) boxes[:, [1, 3]] = np.minimum( height - 1.0, np.maximum(0.0, boxes[:, [1, 3]]) ) return boxes def random_short_side_scale_jitter_list(images, min_size, max_size, boxes=None): """ Perform a spatial short scale jittering on the given images and corresponding boxes. Args: images (list): list of images to perform scale jitter. Dimension is `height` x `width` x `channel`. min_size (int): the minimal size to scale the frames. max_size (int): the maximal size to scale the frames. boxes (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. Returns: (list): the list of scaled images with dimension of `new height` x `new width` x `channel`. (ndarray or None): the scaled boxes with dimension of `num boxes` x 4. """ size = int(round(1.0 / np.random.uniform(1.0 / max_size, 1.0 / min_size))) height = images[0].shape[0] width = images[0].shape[1] if (width <= height and width == size) or ( height <= width and height == size ): return images, boxes new_width = size new_height = size if width < height: new_height = int(math.floor((float(height) / width) * size)) if boxes is not None: boxes = [ proposal * float(new_height) / height for proposal in boxes ] else: new_width = int(math.floor((float(width) / height) * size)) if boxes is not None: boxes = [proposal * float(new_width) / width for proposal in boxes] return ( [ cv2.resize( image, (new_width, new_height), interpolation=cv2.INTER_LINEAR ).astype(np.float32) for image in images ], boxes, ) def scale(size, image): """ Scale the short side of the image to size. Args: size (int): size to scale the image. image (array): image to perform short side scale. Dimension is `height` x `width` x `channel`. Returns: (ndarray): the scaled image with dimension of `height` x `width` x `channel`. """ height = image.shape[0] width = image.shape[1] if (width <= height and width == size) or ( height <= width and height == size ): return image new_width = size new_height = size if width < height: new_height = int(math.floor((float(height) / width) * size)) else: new_width = int(math.floor((float(width) / height) * size)) img = cv2.resize( image, (new_width, new_height), interpolation=cv2.INTER_LINEAR ) return img.astype(np.float32) def scale_boxes(size, boxes, height, width): """ Scale the short side of the box to size. Args: size (int): size to scale the image. boxes (ndarray): bounding boxes to peform scale. The dimension is `num boxes` x 4. height (int): the height of the image. width (int): the width of the image. Returns: boxes (ndarray): scaled bounding boxes. """ if (width <= height and width == size) or ( height <= width and height == size ): return boxes new_width = size new_height = size if width < height: new_height = int(math.floor((float(height) / width) * size)) boxes *= float(new_height) / height else: new_width = int(math.floor((float(width) / height) * size)) boxes *= float(new_width) / width return boxes def horizontal_flip_list(prob, images, order="CHW", boxes=None): """ Horizontally flip the list of image and optional boxes. Args: prob (float): probability to flip. image (list): ilist of images to perform short side scale. Dimension is `height` x `width` x `channel` or `channel` x `height` x `width`. order (str): order of the `height`, `channel` and `width`. boxes (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. Returns: (ndarray): the scaled image with dimension of `height` x `width` x `channel`. (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. """ _, width, _ = images[0].shape if np.random.uniform() < prob: if boxes is not None: boxes = [flip_boxes(proposal, width) for proposal in boxes] if order == "CHW": out_images = [] for image in images: image = np.asarray(image).swapaxes(2, 0) image = image[::-1] out_images.append(image.swapaxes(0, 2)) return out_images, boxes elif order == "HWC": return [cv2.flip(image, 1) for image in images], boxes return images, boxes def spatial_shift_crop_list(size, images, spatial_shift_pos, boxes=None): """ Perform left, center, or right crop of the given list of images. Args: size (int): size to crop. image (list): ilist of images to perform short side scale. Dimension is `height` x `width` x `channel` or `channel` x `height` x `width`. spatial_shift_pos (int): option includes 0 (left), 1 (middle), and 2 (right) crop. boxes (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. Returns: cropped (ndarray): the cropped list of images with dimension of `height` x `width` x `channel`. boxes (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. """ assert spatial_shift_pos in [0, 1, 2] height = images[0].shape[0] width = images[0].shape[1] y_offset = int(math.ceil((height - size) / 2)) x_offset = int(math.ceil((width - size) / 2)) if height > width: if spatial_shift_pos == 0: y_offset = 0 elif spatial_shift_pos == 2: y_offset = height - size else: if spatial_shift_pos == 0: x_offset = 0 elif spatial_shift_pos == 2: x_offset = width - size cropped = [ image[y_offset : y_offset + size, x_offset : x_offset + size, :] for image in images ] assert cropped[0].shape[0] == size, "Image height not cropped properly" assert cropped[0].shape[1] == size, "Image width not cropped properly" if boxes is not None: for i in range(len(boxes)): boxes[i][:, [0, 2]] -= x_offset boxes[i][:, [1, 3]] -= y_offset return cropped, boxes def CHW2HWC(image): """ Transpose the dimension from `channel` x `height` x `width` to `height` x `width` x `channel`. Args: image (array): image to transpose. Returns (array): transposed image. """ return image.transpose([1, 2, 0]) def HWC2CHW(image): """ Transpose the dimension from `height` x `width` x `channel` to `channel` x `height` x `width`. Args: image (array): image to transpose. Returns (array): transposed image. """ return image.transpose([2, 0, 1]) def color_jitter_list( images, img_brightness=0, img_contrast=0, img_saturation=0 ): """ Perform color jitter on the list of images. Args: images (list): list of images to perform color jitter. img_brightness (float): jitter ratio for brightness. img_contrast (float): jitter ratio for contrast. img_saturation (float): jitter ratio for saturation. Returns: images (list): the jittered list of images. """ jitter = [] if img_brightness != 0: jitter.append("brightness") if img_contrast != 0: jitter.append("contrast") if img_saturation != 0: jitter.append("saturation") if len(jitter) > 0: order = np.random.permutation(np.arange(len(jitter))) for idx in range(0, len(jitter)): if jitter[order[idx]] == "brightness": images = brightness_list(img_brightness, images) elif jitter[order[idx]] == "contrast": images = contrast_list(img_contrast, images) elif jitter[order[idx]] == "saturation": images = saturation_list(img_saturation, images) return images def lighting_list(imgs, alphastd, eigval, eigvec, alpha=None): """ Perform AlexNet-style PCA jitter on the given list of images. Args: images (list): list of images to perform lighting jitter. alphastd (float): jitter ratio for PCA jitter. eigval (list): eigenvalues for PCA jitter. eigvec (list[list]): eigenvectors for PCA jitter. Returns: out_images (list): the list of jittered images. """ if alphastd == 0: return imgs # generate alpha1, alpha2, alpha3 alpha = np.random.normal(0, alphastd, size=(1, 3)) eig_vec = np.array(eigvec) eig_val = np.reshape(eigval, (1, 3)) rgb = np.sum( eig_vec * np.repeat(alpha, 3, axis=0) * np.repeat(eig_val, 3, axis=0), axis=1, ) out_images = [] for img in imgs: for idx in range(img.shape[0]): img[idx] = img[idx] + rgb[2 - idx] out_images.append(img) return out_images def color_normalization(image, mean, stddev): """ Perform color normalization on the image with the given mean and stddev. Args: image (array): image to perform color normalization. mean (float): mean value to subtract. stddev (float): stddev to devide. """ # Input image should in format of CHW assert len(mean) == image.shape[0], "channel mean not computed properly" assert len(stddev) == image.shape[0], "channel stddev not computed properly" for idx in range(image.shape[0]): image[idx] = image[idx] - mean[idx] image[idx] = image[idx] / stddev[idx] return image def pad_image(image, pad_size, order="CHW"): """ Pad the given image with the size of pad_size. Args: image (array): image to pad. pad_size (int): size to pad. order (str): order of the `height`, `channel` and `width`. Returns: img (array): padded image. """ if order == "CHW": img = np.pad( image, ((0, 0), (pad_size, pad_size), (pad_size, pad_size)), mode=str("constant"), ) elif order == "HWC": img = np.pad( image, ((pad_size, pad_size), (pad_size, pad_size), (0, 0)), mode=str("constant"), ) return img def horizontal_flip(prob, image, order="CHW"): """ Horizontally flip the image. Args: prob (float): probability to flip. image (array): image to pad. order (str): order of the `height`, `channel` and `width`. Returns: img (array): flipped image. """ assert order in ["CHW", "HWC"], "order {} is not supported".format(order) if np.random.uniform() < prob: if order == "CHW": image = image[:, :, ::-1] elif order == "HWC": image = image[:, ::-1, :] else: raise NotImplementedError("Unknown order {}".format(order)) return image def flip_boxes(boxes, im_width): """ Horizontally flip the boxes. Args: boxes (array): box to flip. im_width (int): width of the image. Returns: boxes_flipped (array): flipped box. """ boxes_flipped = boxes.copy() boxes_flipped[:, 0::4] = im_width - boxes[:, 2::4] - 1 boxes_flipped[:, 2::4] = im_width - boxes[:, 0::4] - 1 return boxes_flipped def crop_boxes(boxes, x_offset, y_offset): """ Crop the boxes given the offsets. Args: boxes (array): boxes to crop. x_offset (int): offset on x. y_offset (int): offset on y. """ boxes[:, [0, 2]] = boxes[:, [0, 2]] - x_offset boxes[:, [1, 3]] = boxes[:, [1, 3]] - y_offset return boxes def random_crop_list(images, size, pad_size=0, order="CHW", boxes=None): """ Perform random crop on a list of images. Args: images (list): list of images to perform random crop. size (int): size to crop. pad_size (int): padding size. order (str): order of the `height`, `channel` and `width`. boxes (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. Returns: cropped (ndarray): the cropped list of images with dimension of `height` x `width` x `channel`. boxes (list): optional. Corresponding boxes to images. Dimension is `num boxes` x 4. """ # explicitly dealing processing per image order to avoid flipping images. if pad_size > 0: images = [ pad_image(pad_size=pad_size, image=image, order=order) for image in images ] # image format should be CHW. if order == "CHW": if images[0].shape[1] == size and images[0].shape[2] == size: return images, boxes height = images[0].shape[1] width = images[0].shape[2] y_offset = 0 if height > size: y_offset = int(np.random.randint(0, height - size)) x_offset = 0 if width > size: x_offset = int(np.random.randint(0, width - size)) cropped = [ image[:, y_offset : y_offset + size, x_offset : x_offset + size] for image in images ] assert cropped[0].shape[1] == size, "Image not cropped properly" assert cropped[0].shape[2] == size, "Image not cropped properly" elif order == "HWC": if images[0].shape[0] == size and images[0].shape[1] == size: return images, boxes height = images[0].shape[0] width = images[0].shape[1] y_offset = 0 if height > size: y_offset = int(np.random.randint(0, height - size)) x_offset = 0 if width > size: x_offset = int(np.random.randint(0, width - size)) cropped = [ image[y_offset : y_offset + size, x_offset : x_offset + size, :] for image in images ] assert cropped[0].shape[0] == size, "Image not cropped properly" assert cropped[0].shape[1] == size, "Image not cropped properly" if boxes is not None: boxes = [crop_boxes(proposal, x_offset, y_offset) for proposal in boxes] return cropped, boxes def center_crop(size, image): """ Perform center crop on input images. Args: size (int): size of the cropped height and width. image (array): the image to perform center crop. """ height = image.shape[0] width = image.shape[1] y_offset = int(math.ceil((height - size) / 2)) x_offset = int(math.ceil((width - size) / 2)) cropped = image[y_offset : y_offset + size, x_offset : x_offset + size, :] assert cropped.shape[0] == size, "Image height not cropped properly" assert cropped.shape[1] == size, "Image width not cropped properly" return cropped # ResNet style scale jittering: randomly select the scale from # [1/max_size, 1/min_size] def random_scale_jitter(image, min_size, max_size): """ Perform ResNet style random scale jittering: randomly select the scale from [1/max_size, 1/min_size]. Args: image (array): image to perform random scale. min_size (int): min size to scale. max_size (int) max size to scale. Returns: image (array): scaled image. """ img_scale = int( round(1.0 / np.random.uniform(1.0 / max_size, 1.0 / min_size)) ) image = scale(img_scale, image) return image def random_scale_jitter_list(images, min_size, max_size): """ Perform ResNet style random scale jittering on a list of image: randomly select the scale from [1/max_size, 1/min_size]. Note that all the image will share the same scale. Args: images (list): list of images to perform random scale. min_size (int): min size to scale. max_size (int) max size to scale. Returns: images (list): list of scaled image. """ img_scale = int( round(1.0 / np.random.uniform(1.0 / max_size, 1.0 / min_size)) ) return [scale(img_scale, image) for image in images] def random_sized_crop(image, size, area_frac=0.08): """ Perform random sized cropping on the given image. Random crop with size 8% - 100% image area and aspect ratio in [3/4, 4/3]. Args: image (array): image to crop. size (int): size to crop. area_frac (float): area of fraction. Returns: (array): cropped image. """ for _ in range(0, 10): height = image.shape[0] width = image.shape[1] area = height * width target_area = np.random.uniform(area_frac, 1.0) * area aspect_ratio = np.random.uniform(3.0 / 4.0, 4.0 / 3.0) w = int(round(math.sqrt(float(target_area) * aspect_ratio))) h = int(round(math.sqrt(float(target_area) / aspect_ratio))) if np.random.uniform() < 0.5: w, h = h, w if h <= height and w <= width: if height == h: y_offset = 0 else: y_offset = np.random.randint(0, height - h) if width == w: x_offset = 0 else: x_offset = np.random.randint(0, width - w) y_offset = int(y_offset) x_offset = int(x_offset) cropped = image[y_offset : y_offset + h, x_offset : x_offset + w, :] assert ( cropped.shape[0] == h and cropped.shape[1] == w ), "Wrong crop size" cropped = cv2.resize( cropped, (size, size), interpolation=cv2.INTER_LINEAR ) return cropped.astype(np.float32) return center_crop(size, scale(size, image)) def lighting(img, alphastd, eigval, eigvec): """ Perform AlexNet-style PCA jitter on the given image. Args: image (array): list of images to perform lighting jitter. alphastd (float): jitter ratio for PCA jitter. eigval (array): eigenvalues for PCA jitter. eigvec (list): eigenvectors for PCA jitter. Returns: img (tensor): the jittered image. """ if alphastd == 0: return img # generate alpha1, alpha2, alpha3. alpha = np.random.normal(0, alphastd, size=(1, 3)) eig_vec = np.array(eigvec) eig_val = np.reshape(eigval, (1, 3)) rgb = np.sum( eig_vec * np.repeat(alpha, 3, axis=0) * np.repeat(eig_val, 3, axis=0), axis=1, ) for idx in range(img.shape[0]): img[idx] = img[idx] + rgb[2 - idx] return img def random_sized_crop_list(images, size, crop_area_fraction=0.08): """ Perform random sized cropping on the given list of images. Random crop with size 8% - 100% image area and aspect ratio in [3/4, 4/3]. Args: images (list): image to crop. size (int): size to crop. area_frac (float): area of fraction. Returns: (list): list of cropped image. """ for _ in range(0, 10): height = images[0].shape[0] width = images[0].shape[1] area = height * width target_area = np.random.uniform(crop_area_fraction, 1.0) * area aspect_ratio = np.random.uniform(3.0 / 4.0, 4.0 / 3.0) w = int(round(math.sqrt(float(target_area) * aspect_ratio))) h = int(round(math.sqrt(float(target_area) / aspect_ratio))) if np.random.uniform() < 0.5: w, h = h, w if h <= height and w <= width: if height == h: y_offset = 0 else: y_offset = np.random.randint(0, height - h) if width == w: x_offset = 0 else: x_offset = np.random.randint(0, width - w) y_offset = int(y_offset) x_offset = int(x_offset) croppsed_images = [] for image in images: cropped = image[ y_offset : y_offset + h, x_offset : x_offset + w, : ] assert ( cropped.shape[0] == h and cropped.shape[1] == w ), "Wrong crop size" cropped = cv2.resize( cropped, (size, size), interpolation=cv2.INTER_LINEAR ) croppsed_images.append(cropped.astype(np.float32)) return croppsed_images return [center_crop(size, scale(size, image)) for image in images] def blend(image1, image2, alpha): return image1 * alpha + image2 * (1 - alpha) def grayscale(image): """ Convert the image to gray scale. Args: image (tensor): image to convert to gray scale. Dimension is `channel` x `height` x `width`. Returns: img_gray (tensor): image in gray scale. """ # R -> 0.299, G -> 0.587, B -> 0.114. img_gray = np.copy(image) gray_channel = 0.299 * image[2] + 0.587 * image[1] + 0.114 * image[0] img_gray[0] = gray_channel img_gray[1] = gray_channel img_gray[2] = gray_channel return img_gray def saturation(var, image): """ Perform color saturation on the given image. Args: var (float): variance. image (array): image to perform color saturation. Returns: (array): image that performed color saturation. """ img_gray = grayscale(image) alpha = 1.0 + np.random.uniform(-var, var) return blend(image, img_gray, alpha) def brightness(var, image): """ Perform color brightness on the given image. Args: var (float): variance. image (array): image to perform color brightness. Returns: (array): image that performed color brightness. """ img_bright = np.zeros(image.shape).astype(image.dtype) alpha = 1.0 + np.random.uniform(-var, var) return blend(image, img_bright, alpha) def contrast(var, image): """ Perform color contrast on the given image. Args: var (float): variance. image (array): image to perform color contrast. Returns: (array): image that performed color contrast. """ img_gray = grayscale(image) img_gray.fill(np.mean(img_gray[0])) alpha = 1.0 + np.random.uniform(-var, var) return blend(image, img_gray, alpha) def saturation_list(var, images): """ Perform color saturation on the list of given images. Args: var (float): variance. images (list): list of images to perform color saturation. Returns: (list): list of images that performed color saturation. """ alpha = 1.0 + np.random.uniform(-var, var) out_images = [] for image in images: img_gray = grayscale(image) out_images.append(blend(image, img_gray, alpha)) return out_images def brightness_list(var, images): """ Perform color brightness on the given list of images. Args: var (float): variance. images (list): list of images to perform color brightness. Returns: (array): list of images that performed color brightness. """ alpha = 1.0 + np.random.uniform(-var, var) out_images = [] for image in images: img_bright = np.zeros(image.shape).astype(image.dtype) out_images.append(blend(image, img_bright, alpha)) return out_images def contrast_list(var, images): """ Perform color contrast on the given list of images. Args: var (float): variance. images (list): list of images to perform color contrast. Returns: (array): image that performed color contrast. """ alpha = 1.0 + np.random.uniform(-var, var) out_images = [] for image in images: img_gray = grayscale(image) img_gray.fill(np.mean(img_gray[0])) out_images.append(blend(image, img_gray, alpha)) return out_images def color_jitter(image, img_brightness=0, img_contrast=0, img_saturation=0): """ Perform color jitter on the given image. Args: image (array): image to perform color jitter. img_brightness (float): jitter ratio for brightness. img_contrast (float): jitter ratio for contrast. img_saturation (float): jitter ratio for saturation. Returns: image (array): the jittered image. """ jitter = [] if img_brightness != 0: jitter.append("brightness") if img_contrast != 0: jitter.append("contrast") if img_saturation != 0: jitter.append("saturation") if len(jitter) > 0: order = np.random.permutation(np.arange(len(jitter))) for idx in range(0, len(jitter)): if jitter[order[idx]] == "brightness": image = brightness(img_brightness, image) elif jitter[order[idx]] == "contrast": image = contrast(img_contrast, image) elif jitter[order[idx]] == "saturation": image = saturation(img_saturation, image) return image def revert_scaled_boxes(size, boxes, img_height, img_width): """ Revert scaled input boxes to match the original image size. Args: size (int): size of the cropped image. boxes (array): shape (num_boxes, 4). img_height (int): height of original image. img_width (int): width of original image. Returns: reverted_boxes (array): boxes scaled back to the original image size. """ scaled_aspect = np.min([img_height, img_width]) scale_ratio = scaled_aspect / size reverted_boxes = boxes * scale_ratio return reverted_boxes
26,476
32.179198
80
py
STTS
STTS-main/MViT/slowfast/datasets/kinetics.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved. import os import random import torch import torch.utils.data import numpy as np from torchvision import transforms import slowfast.utils.logging as logging from slowfast.utils.env import pathmgr from . import decoder as decoder from . import utils as utils from . import video_container as container from .build import DATASET_REGISTRY from .random_erasing import RandomErasing from .transform import create_random_augment logger = logging.get_logger(__name__) @DATASET_REGISTRY.register() class Kinetics(torch.utils.data.Dataset): """ Kinetics video loader. Construct the Kinetics video loader, then sample clips from the videos. For training and validation, a single clip is randomly sampled from every video with random cropping, scaling, and flipping. For testing, multiple clips are uniformaly sampled from every video with uniform cropping. For uniform cropping, we take the left, center, and right crop if the width is larger than height, or take top, center, and bottom crop if the height is larger than the width. """ def __init__(self, cfg, mode, num_retries=10): """ Construct the Kinetics video loader with a given csv file. The format of the csv file is: ``` path_to_video_1 label_1 path_to_video_2 label_2 ... path_to_video_N label_N ``` Args: cfg (CfgNode): configs. mode (string): Options includes `train`, `val`, or `test` mode. For the train and val mode, the data loader will take data from the train or val set, and sample one clip per video. For the test mode, the data loader will take data from test set, and sample multiple clips per video. num_retries (int): number of retries. """ # Only support train, val, and test mode. assert mode in [ "train", "val", "test", ], "Split '{}' not supported for Kinetics".format(mode) self.mode = mode self.cfg = cfg self._video_meta = {} self._num_retries = num_retries # For training or validation mode, one single clip is sampled from every # video. For testing, NUM_ENSEMBLE_VIEWS clips are sampled from every # video. For every clip, NUM_SPATIAL_CROPS is cropped spatially from # the frames. if self.mode in ["train", "val"]: self._num_clips = 1 elif self.mode in ["test"]: self._num_clips = ( cfg.TEST.NUM_ENSEMBLE_VIEWS * cfg.TEST.NUM_SPATIAL_CROPS ) logger.info("Constructing Kinetics {}...".format(mode)) self._construct_loader() self.aug = False self.rand_erase = False self.use_temporal_gradient = False self.temporal_gradient_rate = 0.0 if self.mode == "train" and self.cfg.AUG.ENABLE: self.aug = True if self.cfg.AUG.RE_PROB > 0: self.rand_erase = True def _construct_loader(self): """ Construct the video loader. """ if self.cfg.USE_MINI: path_to_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "mini_{}.csv".format(self.mode) ) else: if self.cfg.TEST.SUBSET == 'full': path_to_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "{}.csv".format(self.mode) ) elif self.cfg.TEST.SUBSET == 'temporal': path_to_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "{}_temporal.csv".format(self.mode) ) else: path_to_file = os.path.join( self.cfg.DATA.PATH_TO_DATA_DIR, "{}_static.csv".format(self.mode) ) assert pathmgr.exists(path_to_file), "{} dir not found".format( path_to_file ) self._path_to_videos = [] self._labels = [] self._spatial_temporal_idx = [] with pathmgr.open(path_to_file, "r") as f: for clip_idx, path_label in enumerate(f.read().splitlines()): assert ( len(path_label.split(self.cfg.DATA.PATH_LABEL_SEPARATOR)) == 2 ) path, label = path_label.split( self.cfg.DATA.PATH_LABEL_SEPARATOR ) for idx in range(self._num_clips): self._path_to_videos.append( os.path.join(self.cfg.DATA.PATH_PREFIX, path) ) self._labels.append(int(label)) self._spatial_temporal_idx.append(idx) self._video_meta[clip_idx * self._num_clips + idx] = {} assert ( len(self._path_to_videos) > 0 ), "Failed to load Kinetics split {} from {}".format( self._split_idx, path_to_file ) logger.info( "Constructing kinetics dataloader (size: {}) from {}".format( len(self._path_to_videos), path_to_file ) ) def __getitem__(self, index): """ Given the video index, return the list of frames, label, and video index if the video can be fetched and decoded successfully, otherwise repeatly find a random video that can be decoded as a replacement. Args: index (int): the video index provided by the pytorch sampler. Returns: frames (tensor): the frames of sampled from the video. The dimension is `channel` x `num frames` x `height` x `width`. label (int): the label of the current video. index (int): if the video provided by pytorch sampler can be decoded, then return the index of the video. If not, return the index of the video replacement that can be decoded. """ # print(index, self._path_to_videos[index]) short_cycle_idx = None # When short cycle is used, input index is a tupple. if isinstance(index, tuple): index, short_cycle_idx = index if self.mode in ["train", "val"]: # -1 indicates random sampling. temporal_sample_index = -1 spatial_sample_index = -1 min_scale = self.cfg.DATA.TRAIN_JITTER_SCALES[0] max_scale = self.cfg.DATA.TRAIN_JITTER_SCALES[1] crop_size = self.cfg.DATA.TRAIN_CROP_SIZE if short_cycle_idx in [0, 1]: crop_size = int( round( self.cfg.MULTIGRID.SHORT_CYCLE_FACTORS[short_cycle_idx] * self.cfg.MULTIGRID.DEFAULT_S ) ) if self.cfg.MULTIGRID.DEFAULT_S > 0: # Decreasing the scale is equivalent to using a larger "span" # in a sampling grid. min_scale = int( round( float(min_scale) * crop_size / self.cfg.MULTIGRID.DEFAULT_S ) ) elif self.mode in ["test"]: temporal_sample_index = ( self._spatial_temporal_idx[index] // self.cfg.TEST.NUM_SPATIAL_CROPS ) # spatial_sample_index is in [0, 1, 2]. Corresponding to left, # center, or right if width is larger than height, and top, middle, # or bottom if height is larger than width. spatial_sample_index = ( ( self._spatial_temporal_idx[index] % self.cfg.TEST.NUM_SPATIAL_CROPS ) if self.cfg.TEST.NUM_SPATIAL_CROPS > 1 else 1 ) min_scale, max_scale, crop_size = ( [self.cfg.DATA.TEST_CROP_SIZE] * 3 if self.cfg.TEST.NUM_SPATIAL_CROPS > 1 else [self.cfg.DATA.TRAIN_JITTER_SCALES[0]] * 2 + [self.cfg.DATA.TEST_CROP_SIZE] ) # The testing is deterministic and no jitter should be performed. # min_scale, max_scale, and crop_size are expect to be the same. assert len({min_scale, max_scale}) == 1 else: raise NotImplementedError( "Does not support {} mode".format(self.mode) ) sampling_rate = utils.get_random_sampling_rate( self.cfg.MULTIGRID.LONG_CYCLE_SAMPLING_RATE, self.cfg.DATA.SAMPLING_RATE, ) # Try to decode and sample a clip from a video. If the video can not be # decoded, repeatly find a random video replacement that can be decoded. for i_try in range(self._num_retries): video_container = None try: video_container = container.get_video_container( self._path_to_videos[index], self.cfg.DATA_LOADER.ENABLE_MULTI_THREAD_DECODE, self.cfg.DATA.DECODING_BACKEND, ) except Exception as e: logger.info( "Failed to load video from {} with error {}".format( self._path_to_videos[index], e ) ) # Select a random video if the current video was not able to access. if video_container is None: logger.warning( "Failed to meta load video idx {} from {}; trial {}".format( index, self._path_to_videos[index], i_try ) ) if self.mode not in ["test"] and i_try > self._num_retries // 2: # let's try another one index = random.randint(0, len(self._path_to_videos) - 1) continue # Decode video. Meta info is used to perform selective decoding. frames = decoder.decode( video_container, sampling_rate, self.cfg.DATA.NUM_FRAMES, temporal_sample_index, self.cfg.TEST.NUM_ENSEMBLE_VIEWS, video_meta=self._video_meta[index], target_fps=self.cfg.DATA.TARGET_FPS, backend=self.cfg.DATA.DECODING_BACKEND, max_spatial_scale=min_scale, use_offset=self.cfg.DATA.USE_OFFSET_SAMPLING, ) # If decoding failed (wrong format, video is too short, and etc), # select another video. if frames is None: logger.warning( "Failed to decode video idx {} from {}; trial {}".format( index, self._path_to_videos[index], i_try ) ) if self.mode not in ["test"] and i_try > self._num_retries // 2: # let's try another one index = random.randint(0, len(self._path_to_videos) - 1) continue if self.aug: if self.cfg.AUG.NUM_SAMPLE > 1: frame_list = [] label_list = [] index_list = [] for _ in range(self.cfg.AUG.NUM_SAMPLE): new_frames = self._aug_frame( frames, spatial_sample_index, min_scale, max_scale, crop_size, ) label = self._labels[index] new_frames = utils.pack_pathway_output( self.cfg, new_frames ) frame_list.append(new_frames) label_list.append(label) index_list.append(index) return frame_list, label_list, index_list, {} else: frames = self._aug_frame( frames, spatial_sample_index, min_scale, max_scale, crop_size, ) else: frames = utils.tensor_normalize( frames, self.cfg.DATA.MEAN, self.cfg.DATA.STD ) # T H W C -> C T H W. frames = frames.permute(3, 0, 1, 2) # Perform data augmentation. frames = utils.spatial_sampling( frames, spatial_idx=spatial_sample_index, min_scale=min_scale, max_scale=max_scale, crop_size=crop_size, random_horizontal_flip=self.cfg.DATA.RANDOM_FLIP, inverse_uniform_sampling=self.cfg.DATA.INV_UNIFORM_SAMPLE, ) label = self._labels[index] frames = utils.pack_pathway_output(self.cfg, frames) return frames, label, index, {} else: raise RuntimeError( "Failed to fetch video after {} retries.".format( self._num_retries ) ) def _aug_frame( self, frames, spatial_sample_index, min_scale, max_scale, crop_size, ): aug_transform = create_random_augment( input_size=(frames.size(1), frames.size(2)), auto_augment=self.cfg.AUG.AA_TYPE, interpolation=self.cfg.AUG.INTERPOLATION, ) # T H W C -> T C H W. frames = frames.permute(0, 3, 1, 2) list_img = self._frame_to_list_img(frames) list_img = aug_transform(list_img) frames = self._list_img_to_frames(list_img) frames = frames.permute(0, 2, 3, 1) frames = utils.tensor_normalize( frames, self.cfg.DATA.MEAN, self.cfg.DATA.STD ) # T H W C -> C T H W. frames = frames.permute(3, 0, 1, 2) # Perform data augmentation. scl, asp = ( self.cfg.DATA.TRAIN_JITTER_SCALES_RELATIVE, self.cfg.DATA.TRAIN_JITTER_ASPECT_RELATIVE, ) relative_scales = ( None if (self.mode not in ["train"] or len(scl) == 0) else scl ) relative_aspect = ( None if (self.mode not in ["train"] or len(asp) == 0) else asp ) frames = utils.spatial_sampling( frames, spatial_idx=spatial_sample_index, min_scale=min_scale, max_scale=max_scale, crop_size=crop_size, random_horizontal_flip=self.cfg.DATA.RANDOM_FLIP, inverse_uniform_sampling=self.cfg.DATA.INV_UNIFORM_SAMPLE, aspect_ratio=relative_aspect, scale=relative_scales, motion_shift=self.cfg.DATA.TRAIN_JITTER_MOTION_SHIFT if self.mode in ["train"] else False, ) if self.rand_erase: erase_transform = RandomErasing( self.cfg.AUG.RE_PROB, mode=self.cfg.AUG.RE_MODE, max_count=self.cfg.AUG.RE_COUNT, num_splits=self.cfg.AUG.RE_COUNT, device="cpu", ) frames = frames.permute(1, 0, 2, 3) frames = erase_transform(frames) frames = frames.permute(1, 0, 2, 3) return frames def _frame_to_list_img(self, frames): img_list = [ transforms.ToPILImage()(frames[i]) for i in range(frames.size(0)) ] return img_list def _list_img_to_frames(self, img_list): img_list = [transforms.ToTensor()(np.array(img)) for img in img_list] return torch.stack(img_list) def __len__(self): """ Returns: (int): the number of videos in the dataset. """ return self.num_videos @property def num_videos(self): """ Returns: (int): the number of videos in the dataset. """ return len(self._path_to_videos)
16,469
37.661972
87
py