mirror of
https://github.com/mozilla/gecko-dev.git
synced 2024-11-29 07:42:04 +00:00
Bug 1131798
- Fix handling of CPU sub-type and rebasing WITHOUT requiring Python 3.3. r=ted
This commit is contained in:
parent
993adae83f
commit
b6225617b0
@ -1,4 +1,5 @@
|
||||
#!/usr/bin/env python
|
||||
#!/usr/bin/env python2.7
|
||||
#
|
||||
# Copyright 2014 Adobe Systems Incorporated. All Rights Reserved.
|
||||
#
|
||||
# Adobe permits you to use, modify, and distribute this file in accordance
|
||||
@ -6,17 +7,18 @@
|
||||
# a copy of the MPL was not distributed with this file, You can obtain one
|
||||
# at http://mozilla.org/MPL/2.0/.
|
||||
#
|
||||
# Create perforce changelist of modules from FTP server
|
||||
# Creates an Adobe Access signed voucher for any executable
|
||||
# Notes: This is currently python2.7 due to mozilla build system requirements
|
||||
|
||||
import io, argparse,pyasn1, bitstring
|
||||
from pyasn1.codec.der import encoder, decoder
|
||||
import argparse, bitstring, pprint, hashlib, os, subprocess, sys, tempfile
|
||||
from pyasn1.codec.der import encoder as der_encoder
|
||||
from pyasn1.type import univ, namedtype, namedval, constraint
|
||||
import hashlib
|
||||
|
||||
|
||||
# CodeSectionDigest ::= SEQUENCE {
|
||||
# offset INTEGER -- section's file offset in the signed binary
|
||||
# digestAlgorithm OBJECT IDENTIFIER -- algorithm identifier for the hash value below. For now only supports SHA256.
|
||||
# digestValue OCTET STRING -- hash value of the TEXT segment.
|
||||
# offset INTEGER -- section's file offset in the signed binary
|
||||
# digestAlgorithm OBJECT IDENTIFIER -- algorithm identifier for the hash value below. For now only supports SHA256.
|
||||
# digestValue OCTET STRING -- hash value of the TEXT segment.
|
||||
# }
|
||||
class CodeSectionDigest(univ.Sequence):
|
||||
componentType = namedtype.NamedTypes(
|
||||
@ -24,6 +26,7 @@ class CodeSectionDigest(univ.Sequence):
|
||||
namedtype.NamedType('digestAlgorithm', univ.ObjectIdentifier()),
|
||||
namedtype.NamedType('digest', univ.OctetString()))
|
||||
|
||||
|
||||
# CodeSegmentDigest ::= SEQUENCE {
|
||||
# offset INTEGER -- TEXT segment's file offset in the signed binary
|
||||
# codeSectionDigests SET OF CodeSectionDigests
|
||||
@ -32,20 +35,22 @@ class CodeSectionDigest(univ.Sequence):
|
||||
class SetOfCodeSectionDigest(univ.SetOf):
|
||||
componentType = CodeSectionDigest()
|
||||
|
||||
|
||||
class CodeSegmentDigest(univ.Sequence):
|
||||
componentType = namedtype.NamedTypes(
|
||||
namedtype.NamedType('offset', univ.Integer()),
|
||||
namedtype.NamedType('codeSectionDigests', SetOfCodeSectionDigest()))
|
||||
|
||||
|
||||
# ArchitectureDigest ::= SEQUENCE {
|
||||
# cpuType ENUMERATED CpuType
|
||||
# cpuSubType ENUMERATED CpuSubType
|
||||
# CodeSegmentDigests SET OF CodeSegmentDigests
|
||||
# }
|
||||
|
||||
class SetOfCodeSegmentDigest(univ.SetOf):
|
||||
componentType = CodeSegmentDigest()
|
||||
|
||||
|
||||
class CPUType(univ.Enumerated):
|
||||
namedValues = namedval.NamedValues(
|
||||
('IMAGE_FILE_MACHINE_I386', 0x14c),
|
||||
@ -55,34 +60,46 @@ class CPUType(univ.Enumerated):
|
||||
constraint.SingleValueConstraint(0x14c, 0x8664)
|
||||
|
||||
|
||||
class CPUSubType(univ.Enumerated):
|
||||
namedValues = namedval.NamedValues(
|
||||
('IMAGE_UNUSED', 0x0),
|
||||
)
|
||||
subtypeSpec = univ.Enumerated.subtypeSpec + \
|
||||
constraint.SingleValueConstraint(0)
|
||||
|
||||
|
||||
class ArchitectureDigest(univ.Sequence):
|
||||
componentType = namedtype.NamedTypes(
|
||||
namedtype.NamedType('cpuType', CPUType()),
|
||||
namedtype.NamedType('cpuSubType', univ.Integer()),
|
||||
namedtype.NamedType('cpuSubType', CPUSubType()),
|
||||
namedtype.NamedType('CodeSegmentDigests', SetOfCodeSegmentDigest())
|
||||
)
|
||||
|
||||
# ApplicationDigest ::= SEQUENCE {
|
||||
# version INTEGER
|
||||
# digests SET OF ArchitectureDigest
|
||||
# }
|
||||
|
||||
# ApplicationDigest ::= SEQUENCE {
|
||||
# version INTEGER
|
||||
# digests SET OF ArchitectureDigest
|
||||
# }
|
||||
class SetOfArchitectureDigest(univ.SetOf):
|
||||
componentType = ArchitectureDigest()
|
||||
|
||||
|
||||
class ApplicationDigest(univ.Sequence):
|
||||
componentType = namedtype.NamedTypes(
|
||||
namedtype.NamedType('version', univ.Integer()),
|
||||
namedtype.NamedType('digests', SetOfArchitectureDigest())
|
||||
)
|
||||
|
||||
def meetsRequirements(items, requirements):
|
||||
|
||||
def meets_requirements(items, requirements):
|
||||
for r in requirements:
|
||||
for n, v in r.items():
|
||||
if n not in items or items[n] != v: return False
|
||||
return True
|
||||
|
||||
|
||||
# return total number of bytes read from items_in excluding leaves
|
||||
def parseItems(stream, items_in, items_out):
|
||||
def parse_items(stream, items_in, items_out):
|
||||
bits_read = 0
|
||||
total_bits_read = 0
|
||||
|
||||
@ -102,7 +119,7 @@ def parseItems(stream, items_in, items_out):
|
||||
requirements = list(filter(lambda x: isinstance(x, dict), item[2]))
|
||||
sub_items = list(filter(lambda x: isinstance(x, tuple), item[2]))
|
||||
|
||||
if not meetsRequirements(items_out, requirements): continue
|
||||
if not meets_requirements(items_out, requirements): continue
|
||||
|
||||
# has sub-items based on length
|
||||
items_out[name] = stream.read(t)
|
||||
@ -113,7 +130,7 @@ def parseItems(stream, items_in, items_out):
|
||||
bit_length = items_out[name] * 8
|
||||
|
||||
if bit_length > 0:
|
||||
sub_read, sub_total_read = parseItems(stream, sub_items, items_out)
|
||||
sub_read, sub_total_read = parse_items(stream, sub_items, items_out)
|
||||
bit_length -= sub_read
|
||||
total_bits_read += sub_total_read
|
||||
|
||||
@ -127,6 +144,7 @@ def parseItems(stream, items_in, items_out):
|
||||
return bits_read, total_bits_read
|
||||
|
||||
|
||||
# TODO: perhaps switch to pefile module when it officially supports python3
|
||||
class SectionHeader:
|
||||
def __init__(self, stream):
|
||||
items = [
|
||||
@ -141,15 +159,115 @@ class SectionHeader:
|
||||
('NumberOfLineNumbers', 'uintle:16'),
|
||||
('Characteristics', 'uintle:32')
|
||||
]
|
||||
self.items = {}
|
||||
_, self.bits_read = parseItems(stream, items, self.items)
|
||||
self.items = dict()
|
||||
self.relocs = dict()
|
||||
|
||||
_, self.bits_read = parse_items(stream, items, self.items)
|
||||
|
||||
self.sectionName = self.items['Name'].decode('utf-8')
|
||||
self.offset = self.items['PointerToRawData']
|
||||
|
||||
COFF_DATA_DIRECTORY_TYPES = [
|
||||
"Export Table",
|
||||
"Import Table",
|
||||
"Resource Table",
|
||||
"Exception Table",
|
||||
"Certificate Tble",
|
||||
"Base Relocation Table",
|
||||
"Debug",
|
||||
"Architecture",
|
||||
"Global Ptr",
|
||||
"TLS Table",
|
||||
"Load Config Table",
|
||||
"Bound Import",
|
||||
"IAT",
|
||||
"Delay Import Descriptor",
|
||||
"CLR Runtime Header",
|
||||
"Reserved",
|
||||
]
|
||||
|
||||
|
||||
def chained_safe_get(obj, names, default=None):
|
||||
if obj is None: return default
|
||||
|
||||
for n in names:
|
||||
if n in obj:
|
||||
obj = obj[n]
|
||||
else:
|
||||
return default
|
||||
|
||||
return obj
|
||||
|
||||
|
||||
class OptionalHeader:
|
||||
def __init__(self, stream, size):
|
||||
self.items = {}
|
||||
items = []
|
||||
|
||||
if size:
|
||||
items += [
|
||||
('Magic', 'uintle:16'),
|
||||
('MajorLinkerVersion', 'uintle:8'),
|
||||
('MinorLinkerVersion', 'uintle:8'),
|
||||
('SizeOfCode', 'uintle:32'),
|
||||
('SizeOfInitializedData', 'uintle:32'),
|
||||
('SizeOfUninitializedData', 'uintle:32'),
|
||||
('AddressOfEntryPoint', 'uintle:32'),
|
||||
('BaseOfCode', 'uintle:32'),
|
||||
]
|
||||
|
||||
_, self.bits_read = parse_items(stream, items, self.items)
|
||||
|
||||
items = []
|
||||
if self.items['Magic'] == 0x10b: # PE32
|
||||
items += [('BaseOfData', 'uintle:32')]
|
||||
|
||||
address_size = 'uintle:64' if self.items['Magic'] == 0x20b else 'uintle:32'
|
||||
|
||||
items += [
|
||||
('ImageBase', address_size),
|
||||
('SectionAlignment', 'uintle:32'),
|
||||
('FileAlignment', 'uintle:32'),
|
||||
('MajorOperatingSystemVersion', 'uintle:16'),
|
||||
('MinorOperatingSystemVersion', 'uintle:16'),
|
||||
('MajorImageVersion', 'uintle:16'),
|
||||
('MinorImageVersion', 'uintle:16'),
|
||||
('MajorSubsystemVersion', 'uintle:16'),
|
||||
('MinorSubsystemVersion', 'uintle:16'),
|
||||
('Win32VersionValue', 'uintle:32'),
|
||||
('SizeOfImage', 'uintle:32'),
|
||||
('SizeOfHeaders', 'uintle:32'),
|
||||
('CheckSum', 'uintle:32'),
|
||||
('Subsystem', 'uintle:16'),
|
||||
('DllCharacteristics', 'uintle:16'),
|
||||
('SizeOfStackReserve', address_size),
|
||||
('SizeOfStackCommit', address_size),
|
||||
('SizeOfHeapReserve', address_size),
|
||||
('SizeOfHeapCommit', address_size),
|
||||
('LoaderFlags', 'uintle:32'),
|
||||
('NumberOfRvaAndSizes', 'uintle:32'),
|
||||
]
|
||||
|
||||
if size > 28:
|
||||
_, bits_read = parse_items(stream, items, self.items)
|
||||
self.bits_read += bits_read
|
||||
|
||||
if 'NumberOfRvaAndSizes' in self.items:
|
||||
index = 0
|
||||
self.items['Data Directories'] = dict()
|
||||
while self.bits_read / 8 < size:
|
||||
d = self.items['Data Directories'][COFF_DATA_DIRECTORY_TYPES[index]] = dict()
|
||||
|
||||
_, bits_read = parse_items(stream, [('VirtualAddress', 'uintle:32'), ('Size', 'uintle:32')], d)
|
||||
self.bits_read += bits_read
|
||||
index += 1
|
||||
|
||||
|
||||
class COFFFileHeader:
|
||||
def __init__(self, stream):
|
||||
self.items = {}
|
||||
self.section_headers = []
|
||||
|
||||
items = [
|
||||
('Machine', 'uintle:16'),
|
||||
('NumberOfSections', 'uintle:16'),
|
||||
@ -159,106 +277,178 @@ class COFFFileHeader:
|
||||
('SizeOfOptionalHeader', 'uintle:16'),
|
||||
('Characteristics', 'uintle:16')
|
||||
]
|
||||
self.items = {}
|
||||
_, self.bits_read = parseItems(stream, items, self.items)
|
||||
_, self.bits_read = parse_items(stream, items, self.items)
|
||||
|
||||
#skip over optional header.
|
||||
if self.items['SizeOfOptionalHeader'] > 0:
|
||||
stream.read(self.items['SizeOfOptionalHeader'] * 8)
|
||||
self.bits_read += self.items['SizeOfOptionalHeader'] * 8
|
||||
self.OptionalHeader = OptionalHeader(stream, self.items['SizeOfOptionalHeader'])
|
||||
self.bits_read += self.OptionalHeader.bits_read
|
||||
|
||||
#start reading section headers
|
||||
numberOfSections = self.items['NumberOfSections']
|
||||
self.codeSectionHeaders = []
|
||||
# start reading section headers
|
||||
num_sections = self.items['NumberOfSections']
|
||||
|
||||
while numberOfSections > 0 :
|
||||
sectionHeader = SectionHeader(stream)
|
||||
if (sectionHeader.items['Characteristics'] & 0x20000000) == 0x20000000:
|
||||
self.codeSectionHeaders.append(sectionHeader)
|
||||
numberOfSections -= 1
|
||||
while num_sections > 0 :
|
||||
section_header = SectionHeader(stream)
|
||||
self.bits_read += section_header.bits_read
|
||||
self.section_headers.append(section_header)
|
||||
num_sections -= 1
|
||||
|
||||
self.codeSectionHeaders.sort(key=lambda header: header.offset)
|
||||
self.section_headers.sort(key=lambda header: header.offset)
|
||||
|
||||
# Read Relocations
|
||||
self.process_relocs(stream)
|
||||
|
||||
def process_relocs(self, stream):
|
||||
reloc_table = chained_safe_get(self.OptionalHeader.items, ['Data Directories', 'Base Relocation Table'])
|
||||
if reloc_table is None: return
|
||||
|
||||
orig_pos = stream.bitpos
|
||||
_, stream.bytepos = self.get_rva_section(reloc_table['VirtualAddress'])
|
||||
end_pos = stream.bitpos + reloc_table['Size'] * 8
|
||||
|
||||
while stream.bitpos < end_pos:
|
||||
page_rva = stream.read('uintle:32')
|
||||
block_size = stream.read('uintle:32')
|
||||
|
||||
for i in range(0, int((block_size - 8) / 2)):
|
||||
data = stream.read('uintle:16')
|
||||
typ = data >> 12
|
||||
offset = data & 0xFFF
|
||||
|
||||
if offset == 0 and i > 0: continue
|
||||
|
||||
assert(typ == 3)
|
||||
|
||||
cur_pos = stream.bitpos
|
||||
sh, value_bytepos = self.get_rva_section(page_rva + offset)
|
||||
stream.bytepos = value_bytepos
|
||||
value = stream.read('uintle:32')
|
||||
|
||||
# remove BaseAddress
|
||||
value -= self.OptionalHeader.items['ImageBase']
|
||||
|
||||
stream.overwrite(bitstring.BitArray(uint=value, length=4 * 8), pos=value_bytepos * 8)
|
||||
stream.pos = cur_pos
|
||||
|
||||
stream.bitpos = orig_pos
|
||||
|
||||
def get_rva_section(self, rva):
|
||||
for sh in self.section_headers:
|
||||
if rva < sh.items['VirtualAddress'] or rva >= sh.items['VirtualAddress'] + sh.items['VirtualSize']:
|
||||
continue
|
||||
|
||||
file_pointer = rva - sh.items['VirtualAddress'] + sh.items['PointerToRawData']
|
||||
return sh, file_pointer
|
||||
|
||||
raise Exception('Could not match RVA to section')
|
||||
|
||||
|
||||
def create_temp_file(suffix=""):
|
||||
fd, path = tempfile.mkstemp(suffix=suffix)
|
||||
os.close(fd)
|
||||
return path
|
||||
|
||||
# TIPS:
|
||||
# How to convert PFX to PEM: openssl pkcs12 -in build/certificates/testPKI/IV.pfx -out build/certificates/testPKI/IV.cert.pem
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description='PE/COFF Parser.')
|
||||
parser = argparse.ArgumentParser(description='PE/COFF Signer')
|
||||
parser.add_argument('-input', required=True, help="File to parse.")
|
||||
parser.add_argument('-output', required=True, help="File to write to.")
|
||||
parser.add_argument('-openssl_path',help="Path to OpenSSL to create signed voucher")
|
||||
parser.add_argument('-signer_cert',help="Path to certificate to use to sign voucher. Must be PEM encoded.")
|
||||
parser.add_argument('-verbose', action='store_true', help="Verbose output.")
|
||||
app_args = parser.parse_args()
|
||||
|
||||
stream = bitstring.ConstBitStream(filename=app_args.input)
|
||||
# to simplify relocation handling we use a mutable BitStream so we can remove
|
||||
# the BaseAddress from each relocation
|
||||
stream = bitstring.BitStream(filename=app_args.input)
|
||||
|
||||
# find the COFF header.
|
||||
# skip forward past the MSDOS stub header to 0x3c.
|
||||
|
||||
stream.bytepos = 0x3c
|
||||
|
||||
# read 4 bytes, this is the file offset of the PE signature.
|
||||
peSignatureOffset = stream.read('uintle:32')
|
||||
stream.bytepos = peSignatureOffset
|
||||
pe_sig_offset = stream.read('uintle:32')
|
||||
stream.bytepos = pe_sig_offset
|
||||
|
||||
#read 4 bytes, make sure it's a PE signature.
|
||||
# read 4 bytes, make sure it's a PE signature.
|
||||
signature = stream.read('uintle:32')
|
||||
if signature != 0x00004550 :
|
||||
return
|
||||
if signature != 0x00004550:
|
||||
raise Exception("Invalid File")
|
||||
|
||||
# after signature is the actual COFF file header.
|
||||
coff_header = COFFFileHeader(stream)
|
||||
|
||||
archDigest = ArchitectureDigest()
|
||||
arch_digest = ArchitectureDigest()
|
||||
if coff_header.items['Machine'] == 0x14c:
|
||||
arch_digest.setComponentByName('cpuType', CPUType('IMAGE_FILE_MACHINE_I386'))
|
||||
elif coff_header.items['Machine'] == 0x8664:
|
||||
arch_digest.setComponentByName('cpuType', CPUType('IMAGE_FILE_MACHINE_AMD64'))
|
||||
|
||||
codeSegmentDigests = SetOfCodeSegmentDigest()
|
||||
codeSegmentIdx = 0
|
||||
arch_digest.setComponentByName('cpuSubType', CPUSubType('IMAGE_UNUSED'))
|
||||
|
||||
#after signature is the actual COFF file header.
|
||||
coffFileHeader = COFFFileHeader(stream)
|
||||
text_section_headers = list(filter(lambda x: (x.items['Characteristics'] & 0x20000000) == 0x20000000, coff_header.section_headers))
|
||||
|
||||
if coffFileHeader.items['Machine'] == 0x14c:
|
||||
archDigest.setComponentByName('cpuType', CPUType('IMAGE_FILE_MACHINE_I386'))
|
||||
elif coffFileHeader.items['Machine'] == 0x8664:
|
||||
archDigest.setComponentByName('cpuType', CPUType('IMAGE_FILE_MACHINE_AMD64'))
|
||||
|
||||
archDigest.setComponentByName('cpuSubType', 0)
|
||||
|
||||
|
||||
for codeSectionHeader in coffFileHeader.codeSectionHeaders:
|
||||
stream.bytepos = codeSectionHeader.offset
|
||||
codeSectionBytes = stream.read('bytes:'+ str(codeSectionHeader.items['SizeOfRawData']))
|
||||
if codeSectionHeader.items['SizeOfRawData'] < codeSectionHeader.items['VirtualSize']:
|
||||
#zero pad up to virtualSize
|
||||
codeSectionBytes += "\0" * (codeSectionHeader.items['VirtualSize']-codeSectionHeader.items['SizeOfRawData'])
|
||||
code_segment_digests = SetOfCodeSegmentDigest()
|
||||
code_segment_idx = 0
|
||||
for code_sect_header in text_section_headers:
|
||||
stream.bytepos = code_sect_header.offset
|
||||
code_sect_bytes = stream.read('bytes:' + str(code_sect_header.items['VirtualSize']))
|
||||
|
||||
digester = hashlib.sha256()
|
||||
digester.update(codeSectionBytes)
|
||||
digester.update(code_sect_bytes)
|
||||
digest = digester.digest()
|
||||
|
||||
codeSectionDigest = CodeSectionDigest()
|
||||
codeSectionDigest.setComponentByName('offset', codeSectionHeader.offset)
|
||||
codeSectionDigest.setComponentByName('digestAlgorithm', univ.ObjectIdentifier('2.16.840.1.101.3.4.2.1'))
|
||||
codeSectionDigest.setComponentByName('digest', univ.OctetString(digest))
|
||||
# with open('segment_' + str(code_sect_header.offset) + ".bin", 'wb') as f:
|
||||
# f.write(code_sect_bytes)
|
||||
|
||||
setOfDigest = SetOfCodeSectionDigest()
|
||||
setOfDigest.setComponentByPosition(0, codeSectionDigest)
|
||||
code_section_digest = CodeSectionDigest()
|
||||
code_section_digest.setComponentByName('offset', code_sect_header.offset)
|
||||
code_section_digest.setComponentByName('digestAlgorithm', univ.ObjectIdentifier('2.16.840.1.101.3.4.2.1'))
|
||||
code_section_digest.setComponentByName('digest', univ.OctetString(digest))
|
||||
|
||||
set_of_digest = SetOfCodeSectionDigest()
|
||||
set_of_digest.setComponentByPosition(0, code_section_digest)
|
||||
|
||||
codeSegmentDigest = CodeSegmentDigest()
|
||||
codeSegmentDigest.setComponentByName('offset', codeSectionHeader.offset)
|
||||
codeSegmentDigest.setComponentByName('codeSectionDigests', setOfDigest)
|
||||
codeSegmentDigest.setComponentByName('offset', code_sect_header.offset)
|
||||
codeSegmentDigest.setComponentByName('codeSectionDigests', set_of_digest)
|
||||
|
||||
codeSegmentDigests.setComponentByPosition(codeSegmentIdx, codeSegmentDigest)
|
||||
codeSegmentIdx += 1
|
||||
code_segment_digests.setComponentByPosition(code_segment_idx, codeSegmentDigest)
|
||||
code_segment_idx += 1
|
||||
|
||||
archDigest.setComponentByName('CodeSegmentDigests', codeSegmentDigests)
|
||||
arch_digest.setComponentByName('CodeSegmentDigests', code_segment_digests)
|
||||
|
||||
setOfArchDigests = SetOfArchitectureDigest()
|
||||
setOfArchDigests.setComponentByPosition(0, archDigest)
|
||||
setOfArchDigests.setComponentByPosition(0, arch_digest)
|
||||
|
||||
appDigest = ApplicationDigest()
|
||||
|
||||
appDigest.setComponentByName('version', 1)
|
||||
appDigest.setComponentByName('digests', setOfArchDigests)
|
||||
|
||||
binaryDigest = encoder.encode(appDigest)
|
||||
binaryDigest = der_encoder.encode(appDigest)
|
||||
|
||||
outFile = open(app_args.output, 'wb')
|
||||
outFile.write(binaryDigest)
|
||||
with open(app_args.output, 'wb') as f:
|
||||
f.write(binaryDigest)
|
||||
|
||||
# sign with openssl if specified
|
||||
if app_args.openssl_path is not None:
|
||||
assert app_args.signer_cert is not None
|
||||
|
||||
out_base, out_ext = os.path.splitext(app_args.output)
|
||||
signed_path = out_base + ".signed" + out_ext
|
||||
|
||||
# http://stackoverflow.com/questions/12507277/how-to-fix-unable-to-write-random-state-in-openssl
|
||||
temp_file = None
|
||||
if sys.platform == "win32" and "RANDFILE" not in os.environ:
|
||||
temp_file = create_temp_file()
|
||||
os.environ["RANDFILE"] = temp_file
|
||||
|
||||
try:
|
||||
subprocess.check_call([app_args.openssl_path, "cms", "-sign", "-nodetach", "-md", "sha256", "-binary", "-in", app_args.output, "-outform", "der", "-out", signed_path, "-signer", app_args.signer_cert], )
|
||||
finally:
|
||||
if temp_file is not None:
|
||||
del os.environ["RANDFILE"]
|
||||
os.unlink(temp_file)
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
|
Loading…
Reference in New Issue
Block a user