2012-07-13 01:07:58 +00:00
import os , os . path , sys
import urllib , zipfile
import shutil , glob , fnmatch
2012-08-02 07:51:51 +00:00
import subprocess , logging , re , shlex
2012-08-08 12:48:06 +00:00
import csv
2012-08-01 10:25:45 +00:00
from hashlib import md5 # pylint: disable-msg=E0611
2012-07-13 01:07:58 +00:00
2012-08-01 13:54:57 +00:00
#class flushout(object):
# def __init__(self, f):
# self.f = f
# def write(self, x):
# self.f.write(x)
# self.f.flush()
#sys.stdout = flushout(sys.stdout)
2012-07-13 01:07:58 +00:00
def download_deps ( mcp_path ) :
bin_path = os . path . normpath ( os . path . join ( mcp_path , ' runtime ' , ' bin ' ) )
ff_path = os . path . normpath ( os . path . join ( bin_path , ' fernflower.jar ' ) )
ret = True
if not os . path . isfile ( ff_path ) :
try :
urllib . urlretrieve ( " http://goo.gl/PnJHp " , ' fernflower.zip ' )
zf = zipfile . ZipFile ( ' fernflower.zip ' )
zf . extract ( ' fernflower.jar ' , bin_path )
zf . close ( )
os . remove ( ' fernflower.zip ' )
print " Downloaded Fernflower successfully "
except :
print " Downloading Fernflower failed download manually from http://goo.gl/PnJHp "
ret = False
2012-08-04 06:51:49 +00:00
for lib in [ ' argo-2.25.jar ' , ' guava-12.0.1.jar ' , ' guava-12.0.1-sources.jar ' , ' asm-all-4.0.jar ' , ' asm-all-4.0-source.jar ' ] :
2012-08-01 10:25:45 +00:00
libF = os . path . join ( mcp_path , ' lib ' )
if not os . path . isdir ( libF ) :
os . makedirs ( libF )
target = os . path . normpath ( os . path . join ( libF , lib ) )
2012-07-13 01:07:58 +00:00
if not os . path . isfile ( target ) :
try :
urllib . urlretrieve ( ' http://cloud.github.com/downloads/cpw/FML/ ' + lib , target )
print ' Downloaded %s successfully ' % lib
except :
2012-08-01 10:25:45 +00:00
print ' Download %s failed, download manually from http://cloud.github.com/downloads/cpw/FML/ %s and place in MCP/lib ' % ( lib , lib )
2012-07-13 01:07:58 +00:00
ret = False
return ret
def pre_decompile ( mcp_dir , fml_dir ) :
2012-08-01 10:25:45 +00:00
file_backup ( os . path . join ( mcp_dir , ' jars ' , ' bin ' ) , ' minecraft.jar ' )
file_backup ( os . path . join ( mcp_dir , ' jars ' ) , ' minecraft_server.jar ' )
def file_backup ( base , file ) :
back_jar = os . path . join ( base , file + ' .backup ' )
src_jar = os . path . join ( base , file )
if not os . path . isfile ( src_jar ) :
return
2012-07-13 01:07:58 +00:00
if os . path . isfile ( back_jar ) :
if os . path . isfile ( src_jar ) :
os . remove ( src_jar )
2012-08-01 10:25:45 +00:00
shutil . copy ( back_jar , src_jar )
else :
shutil . copy ( src_jar , back_jar )
2012-07-13 01:07:58 +00:00
def post_decompile ( mcp_dir , fml_dir ) :
bin_dir = os . path . join ( mcp_dir , ' jars ' , ' bin ' )
back_jar = os . path . join ( bin_dir , ' minecraft.jar.backup ' )
src_jar = os . path . join ( bin_dir , ' minecraft.jar ' )
2012-08-01 10:25:45 +00:00
if not os . path . isfile ( src_jar ) :
return
print ' Stripping META-INF from minecraft.jar '
2012-07-13 01:07:58 +00:00
zip_in = zipfile . ZipFile ( back_jar , mode = ' a ' )
zip_out = zipfile . ZipFile ( src_jar , ' w ' , zipfile . ZIP_DEFLATED )
for i in zip_in . filelist :
if not i . filename . startswith ( ' META-INF ' ) :
c = zip_in . read ( i . filename )
zip_out . writestr ( i . filename , c )
else :
print ' Skipping: %s ' % i . filename
zip_out . close ( )
def reset_logger ( ) :
log = logging . getLogger ( )
while len ( log . handlers ) > 0 :
log . removeHandler ( log . handlers [ 0 ] )
count = 0
def cleanup_source ( path ) :
path = os . path . normpath ( path )
regex_cases_before = re . compile ( r ' ((case|default).+ \ r? \ n) \ r? \ n ' , re . MULTILINE ) #Fixes newline after case before case body
regex_cases_after = re . compile ( r ' \ r? \ n( \ r? \ n[ \ t]+(case|default)) ' , re . MULTILINE ) #Fixes newline after case body before new case
def updatefile ( src_file ) :
global count
tmp_file = src_file + ' .tmp '
count = 0
with open ( src_file , ' r ' ) as fh :
buf = fh . read ( )
def fix_cases ( match ) :
global count
count + = 1
return match . group ( 1 )
buf = regex_cases_before . sub ( fix_cases , buf )
buf = regex_cases_after . sub ( fix_cases , buf )
if count > 0 :
with open ( tmp_file , ' w ' ) as fh :
fh . write ( buf )
shutil . move ( tmp_file , src_file )
for path , _ , filelist in os . walk ( path , followlinks = True ) :
sub_dir = os . path . relpath ( path , path )
for cur_file in fnmatch . filter ( filelist , ' *.java ' ) :
src_file = os . path . normpath ( os . path . join ( path , cur_file ) )
updatefile ( src_file )
def setup_fml ( fml_dir , mcp_dir ) :
sys . path . append ( mcp_dir )
from runtime . decompile import decompile
from runtime . cleanup import cleanup
2012-08-01 10:25:45 +00:00
from runtime . commands import Commands , CLIENT , SERVER
2012-08-03 10:35:38 +00:00
from runtime . mcp import decompile_side , updatemd5_side
2012-07-13 01:07:58 +00:00
src_dir = os . path . join ( mcp_dir , ' src ' )
if os . path . isdir ( src_dir ) :
os . chdir ( mcp_dir )
#cleanup -f
cleanup ( None , False )
reset_logger ( )
os . chdir ( fml_dir )
if os . path . isdir ( src_dir ) :
print ' Please make sure to backup your modified files, and say yes when it asks you to do cleanup. '
sys . exit ( 1 )
#download fernflower/argo/asm/guava
if not download_deps ( mcp_dir ) :
sys . exit ( 1 )
2012-08-01 10:25:45 +00:00
def applyrg_shunt ( self , side , reobf = False , applyrg_real = Commands . applyrg ) :
jars = { CLIENT : self . jarclient , SERVER : self . jarserver }
2012-08-02 22:39:18 +00:00
#print "==================================SHUNT %s============================" % side
#print "Java: %s" % self.cmdjava
#print "Javac: %s" % self.cmdjavac
#print "Jar: %s" % jars[side]
2012-08-01 10:25:45 +00:00
binDir = os . path . join ( fml_dir , ' bin ' )
if not os . path . isdir ( binDir ) :
os . makedirs ( binDir )
#Compile AccessTransformer
2012-08-01 21:29:45 +00:00
forkcmd = ( ' %s -Xlint:-options -deprecation -g -source 1.6 -target 1.6 -classpath " {classpath} " -sourcepath {sourcepath} -d {outpath} {target} ' % self . cmdjavac ) . format (
2012-08-02 07:51:51 +00:00
classpath = os . pathsep . join ( [ ' . ' , os . path . join ( mcp_dir , ' lib ' , ' * ' ) ] ) , sourcepath = os . path . join ( fml_dir , ' common ' ) , outpath = os . path . join ( fml_dir , ' bin ' ) ,
2012-08-04 06:51:49 +00:00
target = os . path . join ( fml_dir , ' common ' , ' cpw ' , ' mods ' , ' fml ' , ' common ' , ' asm ' , ' transformers ' , ' AccessTransformer.java ' ) )
2012-08-03 04:14:06 +00:00
#print forkcmd
2012-08-01 10:25:45 +00:00
self . runcmd ( forkcmd )
2012-08-08 12:48:06 +00:00
#Compile MCPMerger
2012-08-10 03:01:51 +00:00
forkcmd = ( ' %s -Xlint:-options -deprecation -g -source 1.6 -target 1.6 -classpath " {classpath} " -sourcepath " {sourcepath} " -d " {outpath} " " {target} " ' % self . cmdjavac ) . format (
2012-08-08 12:48:06 +00:00
classpath = os . pathsep . join ( [ ' . ' , os . path . join ( mcp_dir , ' lib ' , ' * ' ) ] ) , sourcepath = os . path . join ( fml_dir , ' common ' ) , outpath = os . path . join ( fml_dir , ' bin ' ) ,
target = os . path . join ( fml_dir , ' common ' , ' cpw ' , ' mods ' , ' fml ' , ' common ' , ' asm ' , ' transformers ' , ' MCPMerger.java ' ) )
#print forkcmd
self . runcmd ( forkcmd )
#Run MCPMerger
forkcmd = ( ' %s -classpath " {classpath} " cpw.mods.fml.common.asm.transformers.MCPMerger " {mergecfg} " " {client} " " {server} " ' % self . cmdjava ) . format (
classpath = os . pathsep . join ( [ os . path . join ( mcp_dir , ' lib ' , ' * ' ) , binDir ] ) , mergecfg = os . path . join ( fml_dir , ' mcp_merge.cfg ' ) , client = jars [ CLIENT ] , server = jars [ SERVER ] )
self . runcmd ( forkcmd )
2012-08-01 10:25:45 +00:00
#Run AccessTransformer
forkcmd = ( ' %s -classpath " {classpath} " cpw.mods.fml.common.asm.transformers.AccessTransformer " {jar} " " {fmlconfig} " ' % self . cmdjava ) . format (
classpath = os . pathsep . join ( [ os . path . join ( mcp_dir , ' lib ' , ' * ' ) , binDir ] ) , jar = jars [ side ] , fmlconfig = os . path . join ( fml_dir , ' common ' , ' fml_at.cfg ' ) )
forge_cfg = os . path . join ( mcp_dir , ' forge ' , ' common ' , ' forge_at.cfg ' )
if os . path . isfile ( forge_cfg ) :
forkcmd + = ' " %s " ' % forge_cfg
self . runcmd ( forkcmd )
applyrg_real ( self , side , reobf )
2012-07-13 01:07:58 +00:00
try :
pre_decompile ( mcp_dir , fml_dir )
os . chdir ( mcp_dir )
2012-08-01 10:25:45 +00:00
Commands . applyrg = applyrg_shunt
2012-07-13 01:07:58 +00:00
#decompile -d -n -r
# Conf JAD CSV -r -d -a -n -p -o -l -g
decompile ( None , False , False , True , True , False , True , False , False , False , False )
reset_logger ( )
os . chdir ( fml_dir )
post_decompile ( mcp_dir , fml_dir )
except SystemExit , e :
print ' Decompile Exception: %d ' % e . code
raise e
if not os . path . isdir ( src_dir ) :
print ' Something went wrong, src folder not found at: %s ' % src_dir
sys . exit ( 1 )
#cleanup_source
cleanup_source ( src_dir )
2012-08-01 10:25:45 +00:00
merge_client_server ( mcp_dir )
2012-08-03 10:35:38 +00:00
os . chdir ( mcp_dir )
commands = Commands ( verify = True )
updatemd5_side ( commands , CLIENT )
updatemd5_side ( commands , SERVER )
reset_logger ( )
2012-08-09 01:53:41 +00:00
2012-08-03 10:35:38 +00:00
os . chdir ( fml_dir )
2012-08-01 10:25:45 +00:00
def merge_client_server ( mcp_dir ) :
client = os . path . join ( mcp_dir , ' src ' , ' minecraft ' )
server = os . path . join ( mcp_dir , ' src ' , ' minecraft_server ' )
shared = os . path . join ( mcp_dir , ' src ' , ' common ' )
2012-08-01 13:54:57 +00:00
if not os . path . isdir ( shared ) :
os . makedirs ( shared )
2012-08-01 10:25:45 +00:00
if not os . path . isdir ( client ) or not os . path . isdir ( server ) :
return
if not os . path . isdir ( shared ) :
os . makedirs ( shared )
for path , _ , filelist in os . walk ( client , followlinks = True ) :
for cur_file in filelist :
f_client = os . path . normpath ( os . path . join ( client , path [ len ( client ) + 1 : ] , cur_file ) ) . replace ( os . path . sep , ' / ' )
f_server = os . path . normpath ( os . path . join ( server , path [ len ( client ) + 1 : ] , cur_file ) ) . replace ( os . path . sep , ' / ' )
f_shared = os . path . normpath ( os . path . join ( shared , path [ len ( client ) + 1 : ] , cur_file ) ) . replace ( os . path . sep , ' / ' )
if not os . path . isfile ( f_client ) or not os . path . isfile ( f_server ) :
continue
md5_c = " "
md5_s = " "
with open ( f_client , ' rb ' ) as fh :
md5_c = md5 ( fh . read ( ) ) . hexdigest ( )
with open ( f_server , ' rb ' ) as fh :
md5_s = md5 ( fh . read ( ) ) . hexdigest ( )
if md5_c != md5_s :
continue
new_dir = os . path . join ( shared , path [ len ( client ) + 1 : ] )
if not os . path . isdir ( new_dir ) :
os . makedirs ( new_dir )
shutil . move ( f_client , f_shared )
os . remove ( f_server )
2012-08-08 12:48:06 +00:00
cleanDirs ( server )
cleanDirs ( client )
2012-08-01 13:54:57 +00:00
def apply_fml_patches ( fml_dir , mcp_dir , src_dir , copy_files = True ) :
has_client = os . path . isdir ( os . path . join ( src_dir , ' minecraft ' ) )
has_server = os . path . isdir ( os . path . join ( src_dir , ' minecraft_server ' ) )
2012-07-13 01:07:58 +00:00
2012-08-11 01:25:11 +00:00
#Delete /common/cpw to get rid of the Side/SideOnly classes used in decompilation
cpw_dir = os . path . join ( src_dir , ' common ' , ' cpw ' )
if os . path . isdir ( cpw_dir ) :
shutil . rmtree ( cpw_dir )
2012-07-13 01:07:58 +00:00
#patch files
print ' Applying Forge ModLoader patches '
2012-08-01 13:54:57 +00:00
sys . stdout . flush ( )
2012-07-13 01:07:58 +00:00
if has_client :
if os . path . isdir ( os . path . join ( fml_dir , ' patches ' , ' minecraft ' ) ) :
apply_patches ( mcp_dir , os . path . join ( fml_dir , ' patches ' , ' minecraft ' ) , src_dir )
2012-08-01 13:54:57 +00:00
if os . path . isdir ( os . path . join ( fml_dir , ' patches ' , ' common ' ) ) :
apply_patches ( mcp_dir , os . path . join ( fml_dir , ' patches ' , ' common ' ) , src_dir , ' /common/ ' , ' /minecraft/ ' )
2012-08-04 00:32:23 +00:00
if copy_files and os . path . isdir ( os . path . join ( fml_dir , ' client ' ) ) :
copytree ( os . path . join ( fml_dir , ' client ' ) , os . path . join ( src_dir , ' minecraft ' ) )
2012-07-13 01:07:58 +00:00
#delete argo
2012-08-01 13:54:57 +00:00
if os . path . isdir ( os . path . join ( src_dir , ' minecraft ' , ' argo ' ) ) :
shutil . rmtree ( os . path . join ( src_dir , ' minecraft ' , ' argo ' ) )
2012-07-13 01:07:58 +00:00
if has_server :
if os . path . isdir ( os . path . join ( fml_dir , ' patches ' , ' minecraft_server ' ) ) :
apply_patches ( mcp_dir , os . path . join ( fml_dir , ' patches ' , ' minecraft_server ' ) , src_dir )
2012-08-01 13:54:57 +00:00
if os . path . isdir ( os . path . join ( fml_dir , ' patches ' , ' common ' ) ) :
apply_patches ( mcp_dir , os . path . join ( fml_dir , ' patches ' , ' common ' ) , src_dir , ' /common/ ' , ' /minecraft_server/ ' )
2012-08-04 00:32:23 +00:00
if copy_files and os . path . isdir ( os . path . join ( fml_dir , ' server ' ) ) :
copytree ( os . path . join ( fml_dir , ' server ' ) , os . path . join ( src_dir , ' minecraft_server ' ) )
2012-07-13 01:07:58 +00:00
2012-08-01 13:54:57 +00:00
if os . path . isdir ( os . path . join ( fml_dir , ' patches ' , ' common ' ) ) :
apply_patches ( mcp_dir , os . path . join ( fml_dir , ' patches ' , ' common ' ) , src_dir )
2012-08-04 00:32:23 +00:00
if copy_files and os . path . isdir ( os . path . join ( fml_dir , ' common ' ) ) :
copytree ( os . path . join ( fml_dir , ' common ' ) , os . path . join ( src_dir , ' common ' ) )
2012-07-13 01:07:58 +00:00
def finish_setup_fml ( fml_dir , mcp_dir ) :
sys . path . append ( mcp_dir )
from runtime . updatenames import updatenames
from runtime . updatemd5 import updatemd5
from runtime . updatemcp import updatemcp
os . chdir ( mcp_dir )
updatenames ( None , True )
reset_logger ( )
updatemd5 ( None , True )
reset_logger ( )
os . chdir ( fml_dir )
2012-08-01 13:54:57 +00:00
def apply_patches ( mcp_dir , patch_dir , target_dir , find = None , rep = None ) :
2012-07-13 01:07:58 +00:00
sys . path . append ( mcp_dir )
from runtime . commands import cmdsplit
temp = os . path . abspath ( ' temp.patch ' )
cmd = cmdsplit ( ' patch -p2 -i " %s " ' % temp )
display = True
if os . name == ' nt ' :
applydiff = os . path . abspath ( os . path . join ( mcp_dir , ' runtime ' , ' bin ' , ' applydiff.exe ' ) )
cmd = cmdsplit ( ' " %s " -uf -p2 -i " %s " ' % ( applydiff , temp ) )
display = False
for path , _ , filelist in os . walk ( patch_dir , followlinks = True ) :
for cur_file in fnmatch . filter ( filelist , ' *.patch ' ) :
patch_file = os . path . normpath ( os . path . join ( patch_dir , path [ len ( patch_dir ) + 1 : ] , cur_file ) )
2012-08-01 13:54:57 +00:00
target_file = os . path . join ( target_dir , fix_patch ( patch_file , temp , find , rep ) )
if os . path . isfile ( target_file ) :
if display :
print ' patching file %s ' % os . path . normpath ( target_file )
process = subprocess . Popen ( cmd , cwd = target_dir , bufsize = - 1 )
process . communicate ( )
2012-07-13 01:07:58 +00:00
if os . path . isfile ( temp ) :
os . remove ( temp )
2012-08-01 13:54:57 +00:00
def fix_patch ( in_file , out_file , find = None , rep = None ) :
in_file = os . path . normpath ( in_file )
if out_file is None :
tmp_filename = in_file + ' .tmp '
else :
out_file = os . path . normpath ( out_file )
tmp_file = out_file
dir_name = os . path . dirname ( out_file )
if dir_name :
if not os . path . exists ( dir_name ) :
os . makedirs ( dir_name )
file = ' not found '
with open ( in_file , ' rb ' ) as inpatch :
with open ( tmp_file , ' wb ' ) as outpatch :
for line in inpatch :
line = line . rstrip ( ' \r \n ' )
if line [ : 3 ] in [ ' +++ ' , ' --- ' , ' Onl ' , ' dif ' ] :
2012-08-02 04:38:30 +00:00
if not find == None and not rep == None :
2012-08-01 13:54:57 +00:00
line = line . replace ( ' \\ ' , ' / ' ) . replace ( find , rep ) . replace ( ' / ' , os . sep )
else :
line = line . replace ( ' \\ ' , ' / ' ) . replace ( ' / ' , os . sep )
outpatch . write ( line + os . linesep )
else :
outpatch . write ( line + os . linesep )
if line [ : 3 ] == ' --- ' :
file = line [ line . find ( os . sep , line . find ( os . sep ) + 1 ) + 1 : ]
if out_file is None :
shutil . move ( tmp_file , in_file )
return file
2012-07-13 01:07:58 +00:00
#Taken from: http://stackoverflow.com/questions/7545299/distutil-shutil-copytree
def _mkdir ( newdir ) :
""" works the way a good mkdir should :)
- already exists , silently complete
- regular file in the way , raise an exception
- parent directory ( ies ) does not exist , make them as well
"""
if os . path . isdir ( newdir ) :
pass
elif os . path . isfile ( newdir ) :
raise OSError ( " a file with the same name as the desired " \
" dir, ' %s ' , already exists. " % newdir )
else :
head , tail = os . path . split ( newdir )
if head and not os . path . isdir ( head ) :
_mkdir ( head )
#print "_mkdir %s" % repr(newdir)
if tail :
os . mkdir ( newdir )
#Taken from: http://stackoverflow.com/questions/7545299/distutil-shutil-copytree
def copytree ( src , dst , verbose = 0 , symlinks = False ) :
""" Recursively copy a directory tree using copy2().
The destination directory must not already exist .
If exception ( s ) occur , an Error is raised with a list of reasons .
If the optional symlinks flag is true , symbolic links in the
source tree result in symbolic links in the destination tree ; if
it is false , the contents of the files pointed to by symbolic
links are copied .
XXX Consider this example code rather than the ultimate tool .
"""
if verbose == - 1 :
2012-08-05 03:36:27 +00:00
verbose = len ( os . path . abspath ( dst ) ) + 1
2012-07-13 01:07:58 +00:00
names = os . listdir ( src )
# os.makedirs(dst)
_mkdir ( dst ) # XXX
errors = [ ]
for name in names :
srcname = os . path . join ( src , name )
dstname = os . path . join ( dst , name )
try :
if symlinks and os . path . islink ( srcname ) :
linkto = os . readlink ( srcname )
os . symlink ( linkto , dstname )
elif os . path . isdir ( srcname ) :
copytree ( srcname , dstname , verbose , symlinks )
else :
shutil . copy2 ( srcname , dstname )
if verbose > 0 :
2012-08-05 03:36:27 +00:00
print os . path . abspath ( dstname ) [ verbose : ]
2012-07-13 01:07:58 +00:00
# XXX What about devices, sockets etc.?
except ( IOError , os . error ) , why :
errors . append ( ( srcname , dstname , str ( why ) ) )
# catch the Error from the recursive copytree so that we can
# continue with other files
except Exception , err :
errors . extend ( err . args [ 0 ] )
try :
shutil . copystat ( src , dst )
except WindowsError :
# can't copy file access times on Windows
pass
2012-08-08 12:48:06 +00:00
def cleanDirs ( path ) :
if not os . path . isdir ( path ) :
return
files = os . listdir ( path )
if len ( files ) :
for f in files :
fullpath = os . path . join ( path , f )
if os . path . isdir ( fullpath ) :
cleanDirs ( fullpath )
files = os . listdir ( path )
if len ( files ) == 0 :
os . rmdir ( path )
2012-07-13 01:07:58 +00:00
def merge_tree ( root_src_dir , root_dst_dir ) :
for src_dir , dirs , files in os . walk ( root_src_dir ) :
dst_dir = src_dir . replace ( root_src_dir , root_dst_dir )
if not os . path . exists ( dst_dir ) :
os . mkdir ( dst_dir )
for file_ in files :
src_file = os . path . join ( src_dir , file_ )
dst_file = os . path . join ( dst_dir , file_ )
if os . path . exists ( dst_file ) :
os . remove ( dst_file )
2012-08-01 21:29:45 +00:00
shutil . copy ( src_file , dst_dir )
2012-08-02 07:51:51 +00:00
2012-08-08 12:48:06 +00:00
def setup_mcp ( fml_dir , mcp_dir , dont_gen_conf = True ) :
2012-08-04 06:24:58 +00:00
backup = os . path . join ( mcp_dir , ' runtime ' , ' commands.py.bck ' )
2012-08-02 07:51:51 +00:00
runtime = os . path . join ( mcp_dir , ' runtime ' , ' commands.py ' )
patch = os . path . join ( fml_dir , ' commands.patch ' )
print ' Setting up MCP '
if os . path . isfile ( backup ) :
print ' > Restoring commands.py backup '
if os . path . exists ( runtime ) :
os . remove ( runtime )
shutil . copy ( backup , runtime )
else :
print ' > Backing up commands.py '
shutil . copy ( runtime , backup )
if not os . path . isfile ( patch ) :
return
temp = os . path . abspath ( ' temp.patch ' )
cmd = ' patch -i " %s " ' % temp
windows = os . name == ' nt '
if windows :
applydiff = os . path . abspath ( os . path . join ( mcp_dir , ' runtime ' , ' bin ' , ' applydiff.exe ' ) )
cmd = ' " %s " -uf -i " %s " ' % ( applydiff , temp )
if os . sep == ' \\ ' :
cmd = cmd . replace ( ' \\ ' , ' \\ \\ ' )
cmd = shlex . split ( cmd )
if windows :
print ' Patching file %s ' % os . path . normpath ( runtime )
fix_patch ( patch , temp )
process = subprocess . Popen ( cmd , cwd = os . path . join ( mcp_dir , ' runtime ' ) , bufsize = - 1 )
process . communicate ( )
if os . path . isfile ( temp ) :
2012-08-02 22:39:18 +00:00
os . remove ( temp )
2012-08-03 23:56:00 +00:00
2012-08-08 12:48:06 +00:00
mcp_conf = os . path . join ( mcp_dir , ' conf ' )
mcp_conf_bak = os . path . join ( mcp_dir , ' conf.bak ' )
fml_conf = os . path . join ( fml_dir , ' conf ' )
if not dont_gen_conf :
if os . path . isdir ( mcp_conf_bak ) :
print ' Reverting old conf backup folder '
shutil . rmtree ( mcp_conf )
os . rename ( mcp_conf_bak , mcp_conf )
get_conf_copy ( mcp_dir , fml_dir )
print ' Backing up MCP Conf '
os . rename ( mcp_conf , mcp_conf_bak )
else :
shutil . rmtree ( mcp_conf )
print ' Copying FML conf '
shutil . copytree ( fml_conf , mcp_conf )
2012-08-03 23:56:00 +00:00
#update workspace
print ' Fixing MCP Workspace '
merge_tree ( os . path . join ( fml_dir , ' eclipse ' ) , os . path . join ( mcp_dir , ' eclipse ' ) )
2012-08-11 01:25:11 +00:00
def normaliselines ( in_filename ) :
in_filename = os . path . normpath ( in_filename )
tmp_filename = in_filename + ' .tmp '
with open ( in_filename , ' rb ' ) as in_file :
with open ( tmp_filename , ' wb ' ) as out_file :
out_file . write ( in_file . read ( ) . replace ( ' \r \n ' , ' \n ' ) )
shutil . move ( tmp_filename , in_filename )
2012-08-08 12:48:06 +00:00
def get_conf_copy ( mcp_dir , fml_dir ) :
#Lets grab the files we dont work on
for file in [ ' astyle.cfg ' , ' version.cfg ' , ' patches/minecraft_ff.patch ' , ' patches/minecraft_server_ff.patch ' , ' newids.csv ' ] :
dst_file = os . path . normpath ( os . path . join ( fml_dir , ' conf ' , file ) )
src_file = os . path . normpath ( os . path . join ( mcp_dir , ' conf ' , file ) )
if not os . path . isdir ( os . path . dirname ( dst_file ) ) :
os . makedirs ( os . path . dirname ( dst_file ) )
if os . path . exists ( dst_file ) :
os . remove ( dst_file )
shutil . copy ( src_file , dst_file )
2012-08-11 01:25:11 +00:00
normaliselines ( dst_file )
2012-08-08 12:48:06 +00:00
print ' Grabbing: ' + src_file
ff_server = os . path . normpath ( os . path . join ( fml_dir , ' conf ' , ' patches ' , ' minecraft_server_ff.patch ' ) )
data = [ ]
with open ( ff_server ) as f : data = f . readlines ( ) ;
data = data + [
' diff -r -U 3 minecraft_server \\ net \\ minecraft \\ src \\ ItemMap.java minecraft_server_patched \\ net \\ minecraft \\ src \\ ItemMap.java ' ,
' --- minecraft_server \\ net \\ minecraft \\ src \\ ItemMap.java Wed Aug 01 18:15:37 2012 ' ,
' +++ minecraft_server_patched \\ net \\ minecraft \\ src \\ ItemMap.java Wed Aug 01 18:27:23 2012 ' ,
' @@ -24,7 +24,6 @@ ' ,
' } ' ,
' ' ,
' public static MapData func_77874_a(short p_77874_0_, World p_77874_1_) { ' ,
' - " map_ " + p_77874_0_; ' ,
' MapData var3 = (MapData)p_77874_1_.func_72943_a(MapData.class, " map_ " + p_77874_0_); ' ,
' if(var3 == null) { ' ,
' int var4 = p_77874_1_.func_72841_b( " map " ); '
]
os . remove ( ff_server )
2012-08-11 01:25:11 +00:00
with open ( ff_server , ' wb ' ) as f :
2012-08-08 12:48:06 +00:00
for line in data :
f . write ( line . rstrip ( ' \r \n ' ) + ' \n ' )
common_srg = gen_merged_srg ( mcp_dir , fml_dir )
common_exc = gen_merged_exc ( mcp_dir , fml_dir )
common_map = gen_shared_searge_names ( common_srg , common_exc )
#ToDo use common_map to merge the remaining csvs, client taking precidense, setting the common items to side '2' and editing commands.py in FML to have 'if csv_side == side || csv_side == '2''
gen_merged_csv ( common_map , os . path . join ( mcp_dir , ' conf ' , ' fields.csv ' ) , os . path . join ( fml_dir , ' conf ' , ' fields.csv ' ) )
gen_merged_csv ( common_map , os . path . join ( mcp_dir , ' conf ' , ' methods.csv ' ) , os . path . join ( fml_dir , ' conf ' , ' methods.csv ' ) )
gen_merged_csv ( common_map , os . path . join ( mcp_dir , ' conf ' , ' params.csv ' ) , os . path . join ( fml_dir , ' conf ' , ' params.csv ' ) , main_key = ' param ' )
def gen_merged_srg ( mcp_dir , fml_dir ) :
print ' Generating merged Retroguard data '
srg_client = os . path . join ( mcp_dir , ' conf ' , ' client.srg ' )
srg_server = os . path . join ( mcp_dir , ' conf ' , ' server.srg ' )
if not os . path . isfile ( srg_client ) or not os . path . isfile ( srg_server ) :
print ' Could not find client and server srg files in " %s " ' % mcp_dir
return False
client = { ' PK: ' : { } , ' CL: ' : { } , ' FD: ' : { } , ' MD: ' : { } }
with open ( srg_client , ' r ' ) as fh :
for line in fh :
pts = line . rstrip ( ' \r \n ' ) . split ( ' ' )
if pts [ 0 ] == ' MD: ' :
client [ pts [ 0 ] ] [ pts [ 1 ] + ' ' + pts [ 2 ] ] = pts [ 3 ] + ' ' + pts [ 4 ]
else :
client [ pts [ 0 ] ] [ pts [ 1 ] ] = pts [ 2 ]
server = { ' PK: ' : { } , ' CL: ' : { } , ' FD: ' : { } , ' MD: ' : { } }
with open ( srg_server , ' r ' ) as fh :
for line in fh :
pts = line . rstrip ( ' \r \n ' ) . split ( ' ' )
if pts [ 0 ] == ' MD: ' :
server [ pts [ 0 ] ] [ pts [ 1 ] + ' ' + pts [ 2 ] ] = pts [ 3 ] + ' ' + pts [ 4 ]
else :
server [ pts [ 0 ] ] [ pts [ 1 ] ] = pts [ 2 ]
common = { ' PK: ' : { } , ' CL: ' : { } , ' FD: ' : { } , ' MD: ' : { } }
for type in common :
for key , value in client [ type ] . items ( ) :
if key in server [ type ] :
if value == server [ type ] [ key ] :
client [ type ] . pop ( key )
server [ type ] . pop ( key )
common [ type ] [ key ] = value
for type in common :
for key , value in client [ type ] . items ( ) :
common [ type ] [ key ] = value #+ ' #C'
for type in common :
for key , value in server [ type ] . items ( ) :
common [ type ] [ key ] = value #+ ' #S'
#Print joined retroguard files
2012-08-11 01:25:11 +00:00
with open ( os . path . join ( fml_dir , ' conf ' , ' joined.srg ' ) , ' wb ' ) as f :
2012-08-08 12:48:06 +00:00
for type in [ ' PK: ' , ' CL: ' , ' FD: ' , ' MD: ' ] :
for key in sorted ( common [ type ] ) :
f . write ( ' %s %s %s \n ' % ( type , key , common [ type ] [ key ] ) )
return common
def gen_merged_exc ( mcp_dir , fml_dir ) :
print ' Generating merged MCInjector config '
exc_client = os . path . join ( mcp_dir , ' conf ' , ' client.exc ' )
exc_server = os . path . join ( mcp_dir , ' conf ' , ' server.exc ' )
client = { }
with open ( exc_client , ' r ' ) as fh :
for line in fh :
if not line . startswith ( ' # ' ) :
pts = line . rstrip ( ' \r \n ' ) . split ( ' = ' )
client [ pts [ 0 ] ] = pts [ 1 ]
server = { }
with open ( exc_server , ' r ' ) as fh :
for line in fh :
if not line . startswith ( ' # ' ) :
pts = line . rstrip ( ' \r \n ' ) . split ( ' = ' )
server [ pts [ 0 ] ] = pts [ 1 ]
common = { }
for key , value in client . items ( ) :
if key in server :
if value != server [ key ] :
print ' Error: Exec for shared function does not match client and server: '
print ' Function: ' + key
print ' Client: ' + value
print ' Server: ' + server [ value ]
if value != ' | ' :
common [ key ] = value
client . pop ( key )
server . pop ( key )
else :
if value != ' | ' :
common [ key ] = value
joined = dict ( common . items ( ) + server . items ( ) )
#Print joined mcinjector files
2012-08-11 01:25:11 +00:00
with open ( os . path . join ( fml_dir , ' conf ' , ' joined.exc ' ) , ' wb ' ) as f :
2012-08-08 12:48:06 +00:00
for key in sorted ( joined ) :
f . write ( ' %s = %s \n ' % ( key , joined [ key ] ) )
return common
def gen_shared_searge_names ( common_srg , common_exc ) :
field = re . compile ( r ' field_[0-9]+_[a-zA-Z_]+$ ' )
method = re . compile ( r ' func_[0-9]+_[a-zA-Z_]+ ' )
param = re . compile ( r ' p_[ \ w]+_ \ d+_ ' )
print ' Gathering list of common searge names '
searge = [ ]
for key , value in common_srg [ ' FD: ' ] . items ( ) :
m = field . search ( value )
if not m is None :
if not m . group ( 0 ) in searge :
searge . append ( m . group ( 0 ) )
for key , value in common_srg [ ' MD: ' ] . items ( ) :
m = method . search ( value )
if not m is None and not ' # ' in value :
if not m . group ( 0 ) in searge :
searge . append ( m . group ( 0 ) )
for key , value in common_exc . items ( ) :
m = param . findall ( value )
if not m is None :
for p in m :
if not p in searge :
searge . append ( p )
return searge
def gen_merged_csv ( common_map , in_file , out_file , main_key = ' searge ' ) :
reader = csv . DictReader ( open ( in_file , ' r ' ) )
print ' Generating merged csv for %s ' % os . path . basename ( in_file )
2012-08-11 01:25:11 +00:00
common = [ ]
2012-08-08 12:48:06 +00:00
added = [ ]
for row in reader :
2012-08-11 01:25:11 +00:00
if not row [ main_key ] in added :
row [ ' side ' ] = ' 2 '
common . append ( row )
added . append ( row [ main_key ] )
writer = csv . DictWriter ( open ( out_file , ' wb ' ) , fieldnames = reader . fieldnames , lineterminator = ' \n ' )
2012-08-08 12:48:06 +00:00
writer . writeheader ( )
2012-08-11 01:25:11 +00:00
for row in sorted ( common , key = lambda row : row [ main_key ] ) :
writer . writerow ( row )