SUMMARY = "Updates the NVD CVE database" LICENSE = "MIT" INHIBIT_DEFAULT_DEPS = "1" inherit native deltask do_unpack deltask do_patch deltask do_configure deltask do_compile deltask do_install deltask do_populate_sysroot # CVE database update interval, in seconds. By default: once a day (24*60*60). # Use 0 to force the update # Use a negative value to skip the update CVE_DB_UPDATE_INTERVAL ?= "86400" # Timeout for blocking socket operations, such as the connection attempt. CVE_SOCKET_TIMEOUT ?= "60" NVDCVE_URL ?= "https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-" CVE_DB_TEMP_FILE ?= "${CVE_CHECK_DB_DIR}/temp_nvdcve_1.1.db" python () { if not bb.data.inherits_class("cve-check", d): raise bb.parse.SkipRecipe("Skip recipe when cve-check class is not loaded.") } python do_fetch() { """ Update NVD database with json data feed """ import bb.utils import bb.progress import shutil bb.utils.export_proxies(d) db_file = d.getVar("CVE_CHECK_DB_FILE") db_dir = os.path.dirname(db_file) db_tmp_file = d.getVar("CVE_DB_TEMP_FILE") cleanup_db_download(db_file, db_tmp_file) # The NVD database changes once a day, so no need to update more frequently # Allow the user to force-update try: import time update_interval = int(d.getVar("CVE_DB_UPDATE_INTERVAL")) if update_interval < 0: bb.note("CVE database update skipped") return if time.time() - os.path.getmtime(db_file) < update_interval: return except OSError: pass bb.utils.mkdirhier(db_dir) if os.path.exists(db_file): shutil.copy2(db_file, db_tmp_file) if update_db_file(db_tmp_file, d) == True: # Update downloaded correctly, can swap files shutil.move(db_tmp_file, db_file) else: # Update failed, do not modify the database bb.note("CVE database update failed") os.remove(db_tmp_file) } do_fetch[lockfiles] += "${CVE_CHECK_DB_FILE_LOCK}" do_fetch[file-checksums] = "" do_fetch[vardeps] = "" def cleanup_db_download(db_file, db_tmp_file): """ Cleanup the download space from possible failed downloads """ # Clean up the updates done on the main file # Remove it only if a journal file exists - it means a complete re-download if os.path.exists("{0}-journal".format(db_file)): # If a journal is present the last update might have been interrupted. In that case, # just wipe any leftovers and force the DB to be recreated. os.remove("{0}-journal".format(db_file)) if os.path.exists(db_file): os.remove(db_file) # Clean-up the temporary file downloads, we can remove both journal # and the temporary database if os.path.exists("{0}-journal".format(db_tmp_file)): # If a journal is present the last update might have been interrupted. In that case, # just wipe any leftovers and force the DB to be recreated. os.remove("{0}-journal".format(db_tmp_file)) if os.path.exists(db_tmp_file): os.remove(db_tmp_file) def update_db_file(db_tmp_file, d): """ Update the given database file """ import bb.utils, bb.progress from datetime import date import urllib, gzip, sqlite3 YEAR_START = 2002 cve_socket_timeout = int(d.getVar("CVE_SOCKET_TIMEOUT")) # Connect to database conn = sqlite3.connect(db_tmp_file) initialize_db(conn) with bb.progress.ProgressHandler(d) as ph, open(os.path.join(d.getVar("TMPDIR"), 'cve_check'), 'a') as cve_f: total_years = date.today().year + 1 - YEAR_START for i, year in enumerate(range(YEAR_START, date.today().year + 1)): bb.debug(2, "Updating %d" % year) ph.update((float(i + 1) / total_years) * 100) year_url = (d.getVar('NVDCVE_URL')) + str(year) meta_url = year_url + ".meta" json_url = year_url + ".json.gz" # Retrieve meta last modified date try: response = urllib.request.urlopen(meta_url, timeout=cve_socket_timeout) except urllib.error.URLError as e: cve_f.write('Warning: CVE db update error, Unable to fetch CVE data.\n\n') bb.warn("Failed to fetch CVE data (%s)" % e) import socket result = socket.getaddrinfo("nvd.nist.gov", 443, proto=socket.IPPROTO_TCP) bb.warn("Host IPs are %s" % (", ".join(t[4][0] for t in result))) return False if response: for l in response.read().decode("utf-8").splitlines(): key, value = l.split(":", 1) if key == "lastModifiedDate": last_modified = value break else: bb.warn("Cannot parse CVE metadata, update failed") return False # Compare with current db last modified date cursor = conn.execute("select DATE from META where YEAR = ?", (year,)) meta = cursor.fetchone() cursor.close() if not meta or meta[0] != last_modified: bb.debug(2, "Updating entries") # Clear products table entries corresponding to current year conn.execute("delete from PRODUCTS where ID like ?", ('CVE-%d%%' % year,)).close() # Update db with current year json file try: response = urllib.request.urlopen(json_url, timeout=cve_socket_timeout) if response: update_db(conn, gzip.decompress(response.read()).decode('utf-8')) conn.execute("insert or replace into META values (?, ?)", [year, last_modified]).close() except urllib.error.URLError as e: cve_f.write('Warning: CVE db update error, CVE data is outdated.\n\n') bb.warn("Cannot parse CVE data (%s), update failed" % e.reason) return False else: bb.debug(2, "Already up to date (last modified %s)" % last_modified) # Update success, set the date to cve_check file. if year == date.today().year: cve_f.write('CVE database update : %s\n\n' % date.today()) conn.commit() conn.close() return True def initialize_db(conn): with conn: c = conn.cursor() c.execute("CREATE TABLE IF NOT EXISTS META (YEAR INTEGER UNIQUE, DATE TEXT)") c.execute("CREATE TABLE IF NOT EXISTS NVD (ID TEXT UNIQUE, SUMMARY TEXT, \ SCOREV2 TEXT, SCOREV3 TEXT, MODIFIED INTEGER, VECTOR TEXT)") c.execute("CREATE TABLE IF NOT EXISTS PRODUCTS (ID TEXT, \ VENDOR TEXT, PRODUCT TEXT, VERSION_START TEXT, OPERATOR_START TEXT, \ VERSION_END TEXT, OPERATOR_END TEXT)") c.execute("CREATE INDEX IF NOT EXISTS PRODUCT_ID_IDX on PRODUCTS(ID);") c.close() def parse_node_and_insert(conn, node, cveId): # Parse children node if needed for child in node.get('children', ()): parse_node_and_insert(conn, child, cveId) def cpe_generator(): for cpe in node.get('cpe_match', ()): if not cpe['vulnerable']: return cpe23 = cpe.get('cpe23Uri') if not cpe23: return cpe23 = cpe23.split(':') if len(cpe23) < 6: return vendor = cpe23[3] product = cpe23[4] version = cpe23[5] if cpe23[6] == '*' or cpe23[6] == '-': version_suffix = "" else: version_suffix = "_" + cpe23[6] if version != '*' and version != '-': # Version is defined, this is a '=' match yield [cveId, vendor, product, version + version_suffix, '=', '', ''] elif version == '-': # no version information is available yield [cveId, vendor, product, version, '', '', ''] else: # Parse start version, end version and operators op_start = '' op_end = '' v_start = '' v_end = '' if 'versionStartIncluding' in cpe: op_start = '>=' v_start = cpe['versionStartIncluding'] if 'versionStartExcluding' in cpe: op_start = '>' v_start = cpe['versionStartExcluding'] if 'versionEndIncluding' in cpe: op_end = '<=' v_end = cpe['versionEndIncluding'] if 'versionEndExcluding' in cpe: op_end = '<' v_end = cpe['versionEndExcluding'] if op_start or op_end or v_start or v_end: yield [cveId, vendor, product, v_start, op_start, v_end, op_end] else: # This is no version information, expressed differently. # Save processing by representing as -. yield [cveId, vendor, product, '-', '', '', ''] conn.executemany("insert into PRODUCTS values (?, ?, ?, ?, ?, ?, ?)", cpe_generator()).close() def update_db(conn, jsondata): import json root = json.loads(jsondata) for elt in root['CVE_Items']: if not elt['impact']: continue accessVector = None cveId = elt['cve']['CVE_data_meta']['ID'] cveDesc = elt['cve']['description']['description_data'][0]['value'] date = elt['lastModifiedDate'] try: accessVector = elt['impact']['baseMetricV2']['cvssV2']['accessVector'] cvssv2 = elt['impact']['baseMetricV2']['cvssV2']['baseScore'] except KeyError: cvssv2 = 0.0 try: accessVector = accessVector or elt['impact']['baseMetricV3']['cvssV3']['attackVector'] cvssv3 = elt['impact']['baseMetricV3']['cvssV3']['baseScore'] except KeyError: accessVector = accessVector or "UNKNOWN" cvssv3 = 0.0 conn.execute("insert or replace into NVD values (?, ?, ?, ?, ?, ?)", [cveId, cveDesc, cvssv2, cvssv3, date, accessVector]).close() configurations = elt['configurations']['nodes'] for config in configurations: parse_node_and_insert(conn, config, cveId) do_fetch[nostamp] = "1" EXCLUDE_FROM_WORLD = "1"