1 # This file is part of lyx2lyx
2 # -*- coding: utf-8 -*-
3 # Copyright (C) 2006 José Matos <jamatos@lyx.org>
4 # Copyright (C) 2004-2006 Georg Baum <Georg.Baum@post.rwth-aachen.de>
6 # This program is free software; you can redistribute it and/or
7 # modify it under the terms of the GNU General Public License
8 # as published by the Free Software Foundation; either version 2
9 # of the License, or (at your option) any later version.
11 # This program is distributed in the hope that it will be useful,
12 # but WITHOUT ANY WARRANTY; without even the implied warranty of
13 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 # GNU General Public License for more details.
16 # You should have received a copy of the GNU General Public License
17 # along with this program; if not, write to the Free Software
18 # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
20 """ Convert files to the file format generated by lyx 1.5"""
26 from parser_tools
import find_re
, find_token
, find_token_backwards
, find_token_exact
, find_tokens
, find_end_of
, get_value
, find_beginning_of
, find_nonempty_line
27 from LyX
import get_encoding
30 ####################################################################
31 # Private helper functions
33 def find_end_of_inset(lines
, i
):
34 " Find end of inset, where lines[i] is included."
35 return find_end_of(lines
, i
, "\\begin_inset", "\\end_inset")
37 def find_end_of_layout(lines
, i
):
38 " Find end of layout, where lines[i] is included."
39 return find_end_of(lines
, i
, "\\begin_layout", "\\end_layout")
41 def find_beginning_of_layout(lines
, i
):
42 "Find beginning of layout, where lines[i] is included."
43 return find_beginning_of(lines
, i
, "\\begin_layout", "\\end_layout")
45 # End of helper functions
46 ####################################################################
50 # Notes: Framed/Shaded
53 def revert_framed(document
):
54 "Revert framed notes. "
57 i
= find_tokens(document
.body
, ["\\begin_inset Note Framed", "\\begin_inset Note Shaded"], i
)
61 document
.body
[i
] = "\\begin_inset Note"
69 roman_fonts
= {'default' : 'default', 'ae' : 'ae',
70 'times' : 'times', 'palatino' : 'palatino',
71 'helvet' : 'default', 'avant' : 'default',
72 'newcent' : 'newcent', 'bookman' : 'bookman',
74 sans_fonts
= {'default' : 'default', 'ae' : 'default',
75 'times' : 'default', 'palatino' : 'default',
76 'helvet' : 'helvet', 'avant' : 'avant',
77 'newcent' : 'default', 'bookman' : 'default',
79 typewriter_fonts
= {'default' : 'default', 'ae' : 'default',
80 'times' : 'default', 'palatino' : 'default',
81 'helvet' : 'default', 'avant' : 'default',
82 'newcent' : 'default', 'bookman' : 'default',
83 'pslatex' : 'courier'}
85 def convert_font_settings(document
):
86 " Convert font settings. "
88 i
= find_token_exact(document
.header
, "\\fontscheme", i
)
90 document
.warning("Malformed LyX document: Missing `\\fontscheme'.")
92 font_scheme
= get_value(document
.header
, "\\fontscheme", i
, i
+ 1)
94 document
.warning("Malformed LyX document: Empty `\\fontscheme'.")
95 font_scheme
= 'default'
96 if not font_scheme
in roman_fonts
.keys():
97 document
.warning("Malformed LyX document: Unknown `\\fontscheme' `%s'." % font_scheme
)
98 font_scheme
= 'default'
99 document
.header
[i
:i
+1] = ['\\font_roman %s' % roman_fonts
[font_scheme
],
100 '\\font_sans %s' % sans_fonts
[font_scheme
],
101 '\\font_typewriter %s' % typewriter_fonts
[font_scheme
],
102 '\\font_default_family default',
105 '\\font_sf_scale 100',
106 '\\font_tt_scale 100']
109 def revert_font_settings(document
):
110 " Revert font settings. "
113 fonts
= {'roman' : 'default', 'sans' : 'default', 'typewriter' : 'default'}
114 for family
in 'roman', 'sans', 'typewriter':
115 name
= '\\font_%s' % family
116 i
= find_token_exact(document
.header
, name
, i
)
118 document
.warning("Malformed LyX document: Missing `%s'." % name
)
121 if (insert_line
< 0):
123 fonts
[family
] = get_value(document
.header
, name
, i
, i
+ 1)
124 del document
.header
[i
]
125 i
= find_token_exact(document
.header
, '\\font_default_family', i
)
127 document
.warning("Malformed LyX document: Missing `\\font_default_family'.")
128 font_default_family
= 'default'
130 font_default_family
= get_value(document
.header
, "\\font_default_family", i
, i
+ 1)
131 del document
.header
[i
]
132 i
= find_token_exact(document
.header
, '\\font_sc', i
)
134 document
.warning("Malformed LyX document: Missing `\\font_sc'.")
137 font_sc
= get_value(document
.header
, '\\font_sc', i
, i
+ 1)
138 del document
.header
[i
]
139 if font_sc
!= 'false':
140 document
.warning("Conversion of '\\font_sc' not yet implemented.")
141 i
= find_token_exact(document
.header
, '\\font_osf', i
)
143 document
.warning("Malformed LyX document: Missing `\\font_osf'.")
146 font_osf
= get_value(document
.header
, '\\font_osf', i
, i
+ 1)
147 del document
.header
[i
]
148 i
= find_token_exact(document
.header
, '\\font_sf_scale', i
)
150 document
.warning("Malformed LyX document: Missing `\\font_sf_scale'.")
151 font_sf_scale
= '100'
153 font_sf_scale
= get_value(document
.header
, '\\font_sf_scale', i
, i
+ 1)
154 del document
.header
[i
]
155 if font_sf_scale
!= '100':
156 document
.warning("Conversion of '\\font_sf_scale' not yet implemented.")
157 i
= find_token_exact(document
.header
, '\\font_tt_scale', i
)
159 document
.warning("Malformed LyX document: Missing `\\font_tt_scale'.")
160 font_tt_scale
= '100'
162 font_tt_scale
= get_value(document
.header
, '\\font_tt_scale', i
, i
+ 1)
163 del document
.header
[i
]
164 if font_tt_scale
!= '100':
165 document
.warning("Conversion of '\\font_tt_scale' not yet implemented.")
166 for font_scheme
in roman_fonts
.keys():
167 if (roman_fonts
[font_scheme
] == fonts
['roman'] and
168 sans_fonts
[font_scheme
] == fonts
['sans'] and
169 typewriter_fonts
[font_scheme
] == fonts
['typewriter']):
170 document
.header
.insert(insert_line
, '\\fontscheme %s' % font_scheme
)
171 if font_default_family
!= 'default':
172 document
.preamble
.append('\\renewcommand{\\familydefault}{\\%s}' % font_default_family
)
173 if font_osf
== 'true':
174 document
.warning("Ignoring `\\font_osf = true'")
176 font_scheme
= 'default'
177 document
.header
.insert(insert_line
, '\\fontscheme %s' % font_scheme
)
178 if fonts
['roman'] == 'cmr':
179 document
.preamble
.append('\\renewcommand{\\rmdefault}{cmr}')
180 if font_osf
== 'true':
181 document
.preamble
.append('\\usepackage{eco}')
183 for font
in 'lmodern', 'charter', 'utopia', 'beraserif', 'ccfonts', 'chancery':
184 if fonts
['roman'] == font
:
185 document
.preamble
.append('\\usepackage{%s}' % font
)
186 for font
in 'cmss', 'lmss', 'cmbr':
187 if fonts
['sans'] == font
:
188 document
.preamble
.append('\\renewcommand{\\sfdefault}{%s}' % font
)
189 for font
in 'berasans':
190 if fonts
['sans'] == font
:
191 document
.preamble
.append('\\usepackage{%s}' % font
)
192 for font
in 'cmtt', 'lmtt', 'cmtl':
193 if fonts
['typewriter'] == font
:
194 document
.preamble
.append('\\renewcommand{\\ttdefault}{%s}' % font
)
195 for font
in 'courier', 'beramono', 'luximono':
196 if fonts
['typewriter'] == font
:
197 document
.preamble
.append('\\usepackage{%s}' % font
)
198 if font_default_family
!= 'default':
199 document
.preamble
.append('\\renewcommand{\\familydefault}{\\%s}' % font_default_family
)
200 if font_osf
== 'true':
201 document
.warning("Ignoring `\\font_osf = true'")
204 def revert_booktabs(document
):
205 " We remove the booktabs flag or everything else will become a mess. "
206 re_row
= re
.compile(r
'^<row.*space="[^"]+".*>$')
207 re_tspace
= re
.compile(r
'\s+topspace="[^"]+"')
208 re_bspace
= re
.compile(r
'\s+bottomspace="[^"]+"')
209 re_ispace
= re
.compile(r
'\s+interlinespace="[^"]+"')
212 i
= find_token(document
.body
, "\\begin_inset Tabular", i
)
215 j
= find_end_of_inset(document
.body
, i
+ 1)
217 document
.warning("Malformed LyX document: Could not find end of tabular.")
219 for k
in range(i
, j
):
220 if re
.search('^<features.* booktabs="true".*>$', document
.body
[k
]):
221 document
.warning("Converting 'booktabs' table to normal table.")
222 document
.body
[k
] = document
.body
[k
].replace(' booktabs="true"', '')
223 if re
.search(re_row
, document
.body
[k
]):
224 document
.warning("Removing extra row space.")
225 document
.body
[k
] = re_tspace
.sub('', document
.body
[k
])
226 document
.body
[k
] = re_bspace
.sub('', document
.body
[k
])
227 document
.body
[k
] = re_ispace
.sub('', document
.body
[k
])
231 def convert_multiencoding(document
, forward
):
232 """ Fix files with multiple encodings.
233 Files with an inputencoding of "auto" or "default" and multiple languages
234 where at least two languages have different default encodings are encoded
235 in multiple encodings for file formats < 249. These files are incorrectly
236 read and written (as if the whole file was in the encoding of the main
238 This is not true for files written by CJK-LyX, they are always in the locale
242 - converts from fake unicode values to true unicode if forward is true, and
243 - converts from true unicode values to fake unicode if forward is false.
244 document.encoding must be set to the old value (format 248) in both cases.
246 We do this here and not in LyX.py because it is far easier to do the
247 necessary parsing in modern formats than in ancient ones.
249 inset_types
= ["Foot", "Note"]
250 if document
.cjk_encoding
!= '':
252 encoding_stack
= [document
.encoding
]
254 lang_re
= re
.compile(r
"^\\lang\s(\S+)")
255 inset_re
= re
.compile(r
"^\\begin_inset\s(\S+)")
256 if not forward
: # no need to read file unless we are reverting
257 spec_chars
= read_unicodesymbols()
259 if document
.inputencoding
== "auto" or document
.inputencoding
== "default":
261 while i
< len(document
.body
):
262 result
= lang_re
.match(document
.body
[i
])
264 language
= result
.group(1)
265 if language
== "default":
266 document
.warning("Resetting encoding from %s to %s." % (encoding_stack
[-1], document
.encoding
), 3)
267 encoding_stack
[-1] = document
.encoding
269 from lyx2lyx_lang
import lang
270 document
.warning("Setting encoding from %s to %s." % (encoding_stack
[-1], lang
[language
][3]), 3)
271 encoding_stack
[-1] = lang
[language
][3]
272 elif find_token(document
.body
, "\\begin_layout", i
, i
+ 1) == i
:
273 document
.warning("Adding nested encoding %s." % encoding_stack
[-1], 3)
274 if len(insets
) > 0 and insets
[-1] in inset_types
:
275 from lyx2lyx_lang
import lang
276 encoding_stack
.append(lang
[document
.language
][3])
278 encoding_stack
.append(encoding_stack
[-1])
279 elif find_token(document
.body
, "\\end_layout", i
, i
+ 1) == i
:
280 document
.warning("Removing nested encoding %s." % encoding_stack
[-1], 3)
281 if len(encoding_stack
) == 1:
282 # Don't remove the document encoding from the stack
283 document
.warning("Malformed LyX document: Unexpected `\\end_layout'.")
285 del encoding_stack
[-1]
286 elif find_token(document
.body
, "\\begin_inset", i
, i
+ 1) == i
:
287 inset_result
= inset_re
.match(document
.body
[i
])
289 insets
.append(inset_result
.group(1))
292 elif find_token(document
.body
, "\\end_inset", i
, i
+ 1) == i
:
294 if encoding_stack
[-1] != document
.encoding
:
296 # This line has been incorrectly interpreted as if it was
297 # encoded in 'encoding'.
298 # Convert back to the 8bit string that was in the file.
299 orig
= document
.body
[i
].encode(document
.encoding
)
300 # Convert the 8bit string that was in the file to unicode
301 # with the correct encoding.
302 document
.body
[i
] = orig
.decode(encoding_stack
[-1])
305 # Convert unicode to the 8bit string that will be written
306 # to the file with the correct encoding.
307 orig
= document
.body
[i
].encode(encoding_stack
[-1])
308 # Convert the 8bit string that will be written to the
309 # file to fake unicode with the encoding that will later
310 # be used when writing to the file.
311 document
.body
[i
] = orig
.decode(document
.encoding
)
313 last_char
= document
.body
[i
-1][-1]
314 mod_line
, last_char
= revert_unicode_line(document
, i
, last_char
, insets
, spec_chars
)
315 document
.body
[i
:i
+1] = mod_line
.split('\n')
316 i
+= len(mod_line
.split('\n')) - 1
320 def convert_utf8(document
):
321 " Set document encoding to UTF-8. "
322 convert_multiencoding(document
, True)
323 document
.encoding
= "utf8"
326 def revert_utf8(document
):
327 " Set document encoding to the value corresponding to inputencoding. "
328 i
= find_token(document
.header
, "\\inputencoding", 0)
330 document
.header
.append("\\inputencoding auto")
331 elif get_value(document
.header
, "\\inputencoding", i
) == "utf8":
332 document
.header
[i
] = "\\inputencoding auto"
333 document
.inputencoding
= get_value(document
.header
, "\\inputencoding", 0)
334 document
.encoding
= get_encoding(document
.language
, document
.inputencoding
, 248, document
.cjk_encoding
)
335 convert_multiencoding(document
, False)
338 def read_unicodesymbols():
339 " Read the unicodesymbols list of unicode characters and corresponding commands."
340 pathname
= os
.path
.abspath(os
.path
.dirname(sys
.argv
[0]))
341 fp
= open(os
.path
.join(pathname
.strip('lyx2lyx'), 'unicodesymbols'))
343 for line
in fp
.readlines():
345 line
=line
.replace(' "',' ') # remove all quotation marks with spaces before
346 line
=line
.replace('" ',' ') # remove all quotation marks with spaces after
347 line
=line
.replace(r
'\"','"') # replace \" by " (for characters with diaeresis)
349 # flag1 and flag2 are preamble and other flags
350 [ucs4
,command
,flag1
,flag2
] =line
.split(None,3)
351 spec_chars
[unichr(eval(ucs4
))] = [command
, flag1
, flag2
]
358 def revert_unicode_line(document
, i
, last_char
, insets
, spec_chars
, replacement_character
= '???'):
359 # Define strings to start and end ERT and math insets
360 ert_intro
='\n\n\\begin_inset ERT\nstatus collapsed\n\\begin_layout %s\n\\backslash\n' % document
.default_layout
361 ert_outro
='\n\\end_layout\n\n\\end_inset\n'
362 math_intro
='\n\\begin_inset Formula $'
363 math_outro
='$\n\\end_inset'
366 line
= document
.body
[i
]
367 for character
in line
:
369 # Try to write the character
370 dummy
= character
.encode(document
.encoding
)
371 mod_line
+= character
372 last_char
= character
374 # Try to replace with ERT/math inset
375 if spec_chars
.has_key(character
):
376 command
= spec_chars
[character
][0] # the command to replace unicode
377 flag1
= spec_chars
[character
][1]
378 flag2
= spec_chars
[character
][2]
379 if flag1
.find('combining') > -1 or flag2
.find('combining') > -1:
380 # We have a character that should be combined with the previous
381 command
+= '{' + last_char
+ '}'
382 # Remove the last character. Ignore if it is whitespace
383 if len(last_char
.rstrip()):
384 # last_char was found and is not whitespace
386 mod_line
= mod_line
[:-1]
387 else: # last_char belongs to the last line
388 document
.body
[i
-1] = document
.body
[i
-1][:-1]
390 # The last character was replaced by a command. For now it is
391 # ignored. This could be handled better.
393 if command
[0:2] == '\\\\':
394 if command
[2:12]=='ensuremath':
395 if insets
and insets
[-1] == "ERT":
397 command
= command
.replace('\\\\ensuremath{\\\\', '$\n\\backslash\n')
398 command
= command
.replace('}', '$\n')
399 elif not insets
or insets
[-1] != "Formula":
400 # add a math inset with the replacement character
401 command
= command
.replace('\\\\ensuremath{\\', math_intro
)
402 command
= command
.replace('}', math_outro
)
404 # we are already in a math inset
405 command
= command
.replace('\\\\ensuremath{\\', '')
406 command
= command
.replace('}', '')
408 if insets
and insets
[-1] == "Formula":
409 # avoid putting an ERT in a math; instead put command as text
410 command
= command
.replace('\\\\', '\mathrm{')
411 command
= command
+ '}'
412 elif not insets
or insets
[-1] != "ERT":
413 # add an ERT inset with the replacement character
414 command
= command
.replace('\\\\', ert_intro
)
415 command
= command
+ ert_outro
417 command
= command
.replace('\\\\', '\n\\backslash\n')
418 last_char
= '' # indicate that the character should not be removed
421 # Replace with replacement string
422 mod_line
+= replacement_character
423 return mod_line
, last_char
426 def revert_unicode(document
):
427 '''Transform unicode characters that can not be written using the
428 document encoding to commands according to the unicodesymbols
429 file. Characters that can not be replaced by commands are replaced by
430 an replacement string. Flags other than 'combined' are currently not
432 spec_chars
= read_unicodesymbols()
433 insets
= [] # list of active insets
434 last_char
= '' # to store the previous character
436 # Go through the document to capture all combining characters
438 while i
< len(document
.body
):
439 line
= document
.body
[i
]
441 if line
.find('\\begin_inset') > -1:
442 insets
.append(line
[13:].split()[0])
443 if line
.find('\\end_inset') > -1:
446 # Try to write the line
448 # If all goes well the line is written here
449 dummy
= line
.encode(document
.encoding
)
453 # Error, some character(s) in the line need to be replaced
454 mod_line
, last_char
= revert_unicode_line(document
, i
, last_char
, insets
, spec_chars
)
455 document
.body
[i
:i
+1] = mod_line
.split('\n')
456 i
+= len(mod_line
.split('\n'))
459 def revert_cs_label(document
):
460 " Remove status flag of charstyle label. "
463 i
= find_token(document
.body
, "\\begin_inset CharStyle", i
)
466 # Seach for a line starting 'show_label'
467 # If it is not there, break with a warning message
470 if (document
.body
[i
][:10] == "show_label"):
473 elif (document
.body
[i
][:13] == "\\begin_layout"):
474 document
.warning("Malformed LyX document: Missing 'show_label'.")
481 def convert_bibitem(document
):
483 \bibitem [option]{argument}
487 \begin_inset LatexCommand bibitem
493 This must be called after convert_commandparams.
497 i
= find_token(document
.body
, "\\bibitem", i
)
500 j
= document
.body
[i
].find('[') + 1
501 k
= document
.body
[i
].rfind(']')
502 if j
== 0: # No optional argument found
505 option
= document
.body
[i
][j
:k
]
506 j
= document
.body
[i
].rfind('{') + 1
507 k
= document
.body
[i
].rfind('}')
508 argument
= document
.body
[i
][j
:k
]
509 lines
= ['\\begin_inset LatexCommand bibitem']
511 lines
.append('label "%s"' % option
.replace('"', '\\"'))
512 lines
.append('key "%s"' % argument
.replace('"', '\\"'))
514 lines
.append('\\end_inset')
515 document
.body
[i
:i
+1] = lines
519 commandparams_info
= {
520 # command : [option1, option2, argument]
521 "bibitem" : ["label", "", "key"],
522 "bibtex" : ["options", "btprint", "bibfiles"],
523 "cite" : ["after", "before", "key"],
524 "citet" : ["after", "before", "key"],
525 "citep" : ["after", "before", "key"],
526 "citealt" : ["after", "before", "key"],
527 "citealp" : ["after", "before", "key"],
528 "citeauthor" : ["after", "before", "key"],
529 "citeyear" : ["after", "before", "key"],
530 "citeyearpar" : ["after", "before", "key"],
531 "citet*" : ["after", "before", "key"],
532 "citep*" : ["after", "before", "key"],
533 "citealt*" : ["after", "before", "key"],
534 "citealp*" : ["after", "before", "key"],
535 "citeauthor*" : ["after", "before", "key"],
536 "Citet" : ["after", "before", "key"],
537 "Citep" : ["after", "before", "key"],
538 "Citealt" : ["after", "before", "key"],
539 "Citealp" : ["after", "before", "key"],
540 "Citeauthor" : ["after", "before", "key"],
541 "Citet*" : ["after", "before", "key"],
542 "Citep*" : ["after", "before", "key"],
543 "Citealt*" : ["after", "before", "key"],
544 "Citealp*" : ["after", "before", "key"],
545 "Citeauthor*" : ["after", "before", "key"],
546 "citefield" : ["after", "before", "key"],
547 "citetitle" : ["after", "before", "key"],
548 "cite*" : ["after", "before", "key"],
549 "hfill" : ["", "", ""],
550 "index" : ["", "", "name"],
551 "printindex" : ["", "", "name"],
552 "label" : ["", "", "name"],
553 "eqref" : ["name", "", "reference"],
554 "pageref" : ["name", "", "reference"],
555 "prettyref" : ["name", "", "reference"],
556 "ref" : ["name", "", "reference"],
557 "vpageref" : ["name", "", "reference"],
558 "vref" : ["name", "", "reference"],
559 "tableofcontents" : ["", "", "type"],
560 "htmlurl" : ["name", "", "target"],
561 "url" : ["name", "", "target"]}
564 def convert_commandparams(document
):
567 \begin_inset LatexCommand \cmdname[opt1][opt2]{arg}
572 \begin_inset LatexCommand cmdname
578 name1, name2 and name3 can be different for each command.
580 # \begin_inset LatexCommand bibitem was not the official version (see
581 # convert_bibitem()), but could be read in, so we convert it here, too.
585 i
= find_token(document
.body
, "\\begin_inset LatexCommand", i
)
588 command
= document
.body
[i
][26:].strip()
590 document
.warning("Malformed LyX document: Missing LatexCommand name.")
594 j
= find_token(document
.body
, "\\end_inset", i
+ 1)
596 document
.warning("Malformed document")
598 command
+= "".join(document
.body
[i
+1:j
])
599 document
.body
[i
+1:j
] = []
601 # The following parser is taken from the original InsetCommandParams::scanCommand
607 # Used to handle things like \command[foo[bar]]{foo{bar}}
611 if ((state
== "CMDNAME" and c
== ' ') or
612 (state
== "CMDNAME" and c
== '[') or
613 (state
== "CMDNAME" and c
== '{')):
615 if ((state
== "OPTION" and c
== ']') or
616 (state
== "SECOPTION" and c
== ']') or
617 (state
== "CONTENT" and c
== '}')):
621 nestdepth
= nestdepth
- 1
622 if ((state
== "OPTION" and c
== '[') or
623 (state
== "SECOPTION" and c
== '[') or
624 (state
== "CONTENT" and c
== '{')):
625 nestdepth
= nestdepth
+ 1
626 if state
== "CMDNAME":
628 elif state
== "OPTION":
630 elif state
== "SECOPTION":
632 elif state
== "CONTENT":
637 elif c
== '[' and b
!= ']':
639 nestdepth
= 0 # Just to be sure
640 elif c
== '[' and b
== ']':
642 nestdepth
= 0 # Just to be sure
645 nestdepth
= 0 # Just to be sure
648 # Now we have parsed the command, output the parameters
649 lines
= ["\\begin_inset LatexCommand %s" % name
]
651 if commandparams_info
[name
][0] == "":
652 document
.warning("Ignoring invalid option `%s' of command `%s'." % (option1
, name
))
654 lines
.append('%s "%s"' % (commandparams_info
[name
][0], option1
.replace('"', '\\"')))
656 if commandparams_info
[name
][1] == "":
657 document
.warning("Ignoring invalid second option `%s' of command `%s'." % (option2
, name
))
659 lines
.append('%s "%s"' % (commandparams_info
[name
][1], option2
.replace('"', '\\"')))
661 if commandparams_info
[name
][2] == "":
662 document
.warning("Ignoring invalid argument `%s' of command `%s'." % (argument
, name
))
664 lines
.append('%s "%s"' % (commandparams_info
[name
][2], argument
.replace('"', '\\"')))
665 document
.body
[i
:i
+1] = lines
669 def revert_commandparams(document
):
670 regex
= re
.compile(r
'(\S+)\s+(.+)')
673 i
= find_token(document
.body
, "\\begin_inset LatexCommand", i
)
676 name
= document
.body
[i
].split()[2]
677 j
= find_end_of_inset(document
.body
, i
+ 1)
682 for k
in range(i
+ 1, j
):
683 match
= re
.match(regex
, document
.body
[k
])
685 pname
= match
.group(1)
686 pvalue
= match
.group(2)
687 if pname
== "preview":
688 preview_line
= document
.body
[k
]
689 elif (commandparams_info
[name
][0] != "" and
690 pname
== commandparams_info
[name
][0]):
691 option1
= pvalue
.strip('"').replace('\\"', '"')
692 elif (commandparams_info
[name
][1] != "" and
693 pname
== commandparams_info
[name
][1]):
694 option2
= pvalue
.strip('"').replace('\\"', '"')
695 elif (commandparams_info
[name
][2] != "" and
696 pname
== commandparams_info
[name
][2]):
697 argument
= pvalue
.strip('"').replace('\\"', '"')
698 elif document
.body
[k
].strip() != "":
699 document
.warning("Ignoring unknown contents `%s' in command inset %s." % (document
.body
[k
], name
))
700 if name
== "bibitem":
702 lines
= ["\\bibitem {%s}" % argument
]
704 lines
= ["\\bibitem [%s]{%s}" % (option1
, argument
)]
708 lines
= ["\\begin_inset LatexCommand \\%s{%s}" % (name
, argument
)]
710 lines
= ["\\begin_inset LatexCommand \\%s[][%s]{%s}" % (name
, option2
, argument
)]
713 lines
= ["\\begin_inset LatexCommand \\%s[%s]{%s}" % (name
, option1
, argument
)]
715 lines
= ["\\begin_inset LatexCommand \\%s[%s][%s]{%s}" % (name
, option1
, option2
, argument
)]
716 if name
!= "bibitem":
717 if preview_line
!= "":
718 lines
.append(preview_line
)
720 lines
.append('\\end_inset')
721 document
.body
[i
:j
+1] = lines
725 def revert_nomenclature(document
):
726 " Convert nomenclature entry to ERT. "
727 regex
= re
.compile(r
'(\S+)\s+(.+)')
731 i
= find_token(document
.body
, "\\begin_inset LatexCommand nomenclature", i
)
735 j
= find_end_of_inset(document
.body
, i
+ 1)
740 for k
in range(i
+ 1, j
):
741 match
= re
.match(regex
, document
.body
[k
])
743 name
= match
.group(1)
744 value
= match
.group(2)
745 if name
== "preview":
746 preview_line
= document
.body
[k
]
747 elif name
== "symbol":
748 symbol
= value
.strip('"').replace('\\"', '"')
749 elif name
== "description":
750 description
= value
.strip('"').replace('\\"', '"')
751 elif name
== "prefix":
752 prefix
= value
.strip('"').replace('\\"', '"')
753 elif document
.body
[k
].strip() != "":
754 document
.warning("Ignoring unknown contents `%s' in nomenclature inset." % document
.body
[k
])
756 command
= 'nomenclature{%s}{%s}' % (symbol
, description
)
758 command
= 'nomenclature[%s]{%s}{%s}' % (prefix
, symbol
, description
)
759 document
.body
[i
:j
+1] = ['\\begin_inset ERT',
762 '\\begin_layout %s' % document
.default_layout
,
771 if use_nomencl
and find_token(document
.preamble
, '\\usepackage{nomencl}[2005/09/22]', 0) == -1:
772 document
.preamble
.append('\\usepackage{nomencl}[2005/09/22]')
773 document
.preamble
.append('\\makenomenclature')
776 def revert_printnomenclature(document
):
777 " Convert printnomenclature to ERT. "
778 regex
= re
.compile(r
'(\S+)\s+(.+)')
782 i
= find_token(document
.body
, "\\begin_inset LatexCommand printnomenclature", i
)
786 j
= find_end_of_inset(document
.body
, i
+ 1)
789 for k
in range(i
+ 1, j
):
790 match
= re
.match(regex
, document
.body
[k
])
792 name
= match
.group(1)
793 value
= match
.group(2)
794 if name
== "preview":
795 preview_line
= document
.body
[k
]
796 elif name
== "labelwidth":
797 labelwidth
= value
.strip('"').replace('\\"', '"')
798 elif document
.body
[k
].strip() != "":
799 document
.warning("Ignoring unknown contents `%s' in printnomenclature inset." % document
.body
[k
])
801 command
= 'nomenclature{}'
803 command
= 'nomenclature[%s]' % labelwidth
804 document
.body
[i
:j
+1] = ['\\begin_inset ERT',
807 '\\begin_layout %s' % document
.default_layout
,
816 if use_nomencl
and find_token(document
.preamble
, '\\usepackage{nomencl}[2005/09/22]', 0) == -1:
817 document
.preamble
.append('\\usepackage{nomencl}[2005/09/22]')
818 document
.preamble
.append('\\makenomenclature')
821 def convert_esint(document
):
822 " Add \\use_esint setting to header. "
823 i
= find_token(document
.header
, "\\cite_engine", 0)
825 document
.warning("Malformed LyX document: Missing `\\cite_engine'.")
827 # 0 is off, 1 is auto, 2 is on.
828 document
.header
.insert(i
, '\\use_esint 0')
831 def revert_esint(document
):
832 " Remove \\use_esint setting from header. "
833 i
= find_token(document
.header
, "\\use_esint", 0)
835 document
.warning("Malformed LyX document: Missing `\\use_esint'.")
837 use_esint
= document
.header
[i
].split()[1]
838 del document
.header
[i
]
839 # 0 is off, 1 is auto, 2 is on.
841 document
.preamble
.append('\\usepackage{esint}')
844 def revert_clearpage(document
):
848 i
= find_token(document
.body
, "\\clearpage", i
)
851 document
.body
[i
:i
+1] = ['\\begin_inset ERT',
854 '\\begin_layout %s' % document
.default_layout
,
865 def revert_cleardoublepage(document
):
866 " cleardoublepage -> ERT "
869 i
= find_token(document
.body
, "\\cleardoublepage", i
)
872 document
.body
[i
:i
+1] = ['\\begin_inset ERT',
875 '\\begin_layout %s' % document
.default_layout
,
886 def convert_lyxline(document
):
887 " remove fontsize commands for \lyxline "
888 # The problematic is: The old \lyxline definition doesn't handle the fontsize
889 # to change the line thickness. The new definiton does this so that imported
890 # \lyxlines would have a different line thickness. The eventual fontsize command
891 # before \lyxline is therefore removed to get the same output.
892 fontsizes
= ["tiny", "scriptsize", "footnotesize", "small", "normalsize",
893 "large", "Large", "LARGE", "huge", "Huge"]
894 for n
in range(0, len(fontsizes
)):
897 while i
< len(document
.body
):
898 i
= find_token(document
.body
, "\\size " + fontsizes
[n
], i
)
899 k
= find_token(document
.body
, "\\lyxline", i
)
900 # the corresponding fontsize command is always 2 lines before the \lyxline
901 if (i
!= -1 and k
== i
+2):
902 document
.body
[i
:i
+1] = []
908 def revert_encodings(document
):
909 " Set new encodings to auto. "
910 encodings
= ["8859-6", "8859-8", "cp437", "cp437de", "cp850", "cp852",
911 "cp855", "cp858", "cp862", "cp865", "cp866", "cp1250",
912 "cp1252", "cp1256", "cp1257", "latin10", "pt254", "tis620-0"]
913 i
= find_token(document
.header
, "\\inputencoding", 0)
915 document
.header
.append("\\inputencoding auto")
917 inputenc
= get_value(document
.header
, "\\inputencoding", i
)
918 if inputenc
in encodings
:
919 document
.header
[i
] = "\\inputencoding auto"
920 document
.inputencoding
= get_value(document
.header
, "\\inputencoding", 0)
923 def convert_caption(document
):
924 " Convert caption layouts to caption insets. "
927 i
= find_token(document
.body
, "\\begin_layout Caption", i
)
930 j
= find_end_of_layout(document
.body
, i
)
932 document
.warning("Malformed LyX document: Missing `\\end_layout'.")
935 document
.body
[j
:j
] = ["\\end_layout", "", "\\end_inset", "", ""]
936 document
.body
[i
:i
+1] = ["\\begin_layout %s" % document
.default_layout
,
937 "\\begin_inset Caption", "",
938 "\\begin_layout %s" % document
.default_layout
]
942 def revert_caption(document
):
943 " Convert caption insets to caption layouts. "
944 " This assumes that the text class has a caption style. "
947 i
= find_token(document
.body
, "\\begin_inset Caption", i
)
951 # We either need to delete the previous \begin_layout line, or we
952 # need to end the previous layout if this inset is not in the first
953 # position of the paragraph.
954 layout_before
= find_token_backwards(document
.body
, "\\begin_layout", i
)
955 if layout_before
== -1:
956 document
.warning("Malformed LyX document: Missing `\\begin_layout'.")
958 layout_line
= document
.body
[layout_before
]
959 del_layout_before
= True
960 l
= layout_before
+ 1
962 if document
.body
[l
] != "":
963 del_layout_before
= False
966 if del_layout_before
:
967 del document
.body
[layout_before
:i
]
970 document
.body
[i
:i
] = ["\\end_layout", ""]
973 # Find start of layout in the inset and end of inset
974 j
= find_token(document
.body
, "\\begin_layout", i
)
976 document
.warning("Malformed LyX document: Missing `\\begin_layout'.")
978 k
= find_end_of_inset(document
.body
, i
)
980 document
.warning("Malformed LyX document: Missing `\\end_inset'.")
983 # We either need to delete the following \end_layout line, or we need
984 # to restart the old layout if this inset is not at the paragraph end.
985 layout_after
= find_token(document
.body
, "\\end_layout", k
)
986 if layout_after
== -1:
987 document
.warning("Malformed LyX document: Missing `\\end_layout'.")
989 del_layout_after
= True
991 while l
< layout_after
:
992 if document
.body
[l
] != "":
993 del_layout_after
= False
997 del document
.body
[k
+1:layout_after
+1]
999 document
.body
[k
+1:k
+1] = [layout_line
, ""]
1001 # delete \begin_layout and \end_inset and replace \begin_inset with
1002 # "\begin_layout Caption". This works because we can only have one
1003 # paragraph in the caption inset: The old \end_layout will be recycled.
1004 del document
.body
[k
]
1005 if document
.body
[k
] == "":
1006 del document
.body
[k
]
1007 del document
.body
[j
]
1008 if document
.body
[j
] == "":
1009 del document
.body
[j
]
1010 document
.body
[i
] = "\\begin_layout Caption"
1011 if document
.body
[i
+1] == "":
1012 del document
.body
[i
+1]
1016 # Accents of InsetLaTeXAccent
1018 "`" : u
'\u0300', # grave
1019 "'" : u
'\u0301', # acute
1020 "^" : u
'\u0302', # circumflex
1021 "~" : u
'\u0303', # tilde
1022 "=" : u
'\u0304', # macron
1023 "u" : u
'\u0306', # breve
1024 "." : u
'\u0307', # dot above
1025 "\"": u
'\u0308', # diaeresis
1026 "r" : u
'\u030a', # ring above
1027 "H" : u
'\u030b', # double acute
1028 "v" : u
'\u030c', # caron
1029 "b" : u
'\u0320', # minus sign below
1030 "d" : u
'\u0323', # dot below
1031 "c" : u
'\u0327', # cedilla
1032 "k" : u
'\u0328', # ogonek
1033 "t" : u
'\u0361' # tie. This is special: It spans two characters, but
1034 # only one is given as argument, so we don't need to
1035 # treat it differently.
1039 # special accents of InsetLaTeXAccent without argument
1040 special_accent_map
= {
1041 'i' : u
'\u0131', # dotless i
1042 'j' : u
'\u0237', # dotless j
1043 'l' : u
'\u0142', # l with stroke
1044 'L' : u
'\u0141' # L with stroke
1048 # special accent arguments of InsetLaTeXAccent
1050 '\\i' : u
'\u0131', # dotless i
1051 '\\j' : u
'\u0237' # dotless j
1055 def _convert_accent(accent
, accented_char
):
1057 char
= accented_char
1059 if type in special_accent_map
:
1060 return special_accent_map
[type]
1061 # a missing char is treated as space by LyX
1063 elif type == 'q' and char
in ['t', 'd', 'l', 'L']:
1064 # Special caron, only used with t, d, l and L.
1065 # It is not in the map because we convert it to the same unicode
1066 # character as the normal caron: \q{} is only defined if babel with
1067 # the czech or slovak language is used, and the normal caron
1068 # produces the correct output if the T1 font encoding is used.
1069 # For the same reason we never convert to \q{} in the other direction.
1071 elif char
in accented_map
:
1072 char
= accented_map
[char
]
1073 elif (len(char
) > 1):
1074 # We can only convert accents on a single char
1076 a
= accent_map
.get(type)
1078 return unicodedata
.normalize("NFC", "%s%s" % (char
, a
))
1082 def convert_ertbackslash(body
, i
, ert
, default_layout
):
1083 r
""" -------------------------------------------------------------------------------------------
1084 Convert backslashes and '\n' into valid ERT code, append the converted
1085 text to body[i] and return the (maybe incremented) line index i"""
1089 body
[i
] = body
[i
] + '\\backslash '
1093 body
[i
+1:i
+1] = ['\\end_layout', '', '\\begin_layout %s' % default_layout
, '']
1096 body
[i
] = body
[i
] + c
1100 def convert_accent(document
):
1101 # The following forms are supported by LyX:
1102 # '\i \"{a}' (standard form, as written by LyX)
1103 # '\i \"{}' (standard form, as written by LyX if the accented char is a space)
1104 # '\i \"{ }' (also accepted if the accented char is a space)
1105 # '\i \" a' (also accepted)
1106 # '\i \"' (also accepted)
1107 re_wholeinset
= re
.compile(r
'^(.*)(\\i\s+)(.*)$')
1108 re_contents
= re
.compile(r
'^([^\s{]+)(.*)$')
1109 re_accentedcontents
= re
.compile(r
'^\s*{?([^{}]*)}?\s*$')
1112 i
= find_re(document
.body
, re_wholeinset
, i
)
1115 match
= re_wholeinset
.match(document
.body
[i
])
1116 prefix
= match
.group(1)
1117 contents
= match
.group(3).strip()
1118 match
= re_contents
.match(contents
)
1120 # Strip first char (always \)
1121 accent
= match
.group(1)[1:]
1122 accented_contents
= match
.group(2).strip()
1123 match
= re_accentedcontents
.match(accented_contents
)
1124 accented_char
= match
.group(1)
1125 converted
= _convert_accent(accent
, accented_char
)
1127 # Normalize contents
1128 contents
= '%s{%s}' % (accent
, accented_char
),
1130 document
.body
[i
] = '%s%s' % (prefix
, converted
)
1133 document
.warning("Converting unknown InsetLaTeXAccent `\\i %s' to ERT." % contents
)
1134 document
.body
[i
] = prefix
1135 document
.body
[i
+1:i
+1] = ['\\begin_inset ERT',
1138 '\\begin_layout %s' % document
.default_layout
,
1142 i
= convert_ertbackslash(document
.body
, i
+ 7,
1144 document
.default_layout
)
1145 document
.body
[i
+1:i
+1] = ['\\end_layout',
1151 def revert_accent(document
):
1152 inverse_accent_map
= {}
1153 for k
in accent_map
:
1154 inverse_accent_map
[accent_map
[k
]] = k
1155 inverse_special_accent_map
= {}
1156 for k
in special_accent_map
:
1157 inverse_special_accent_map
[special_accent_map
[k
]] = k
1158 inverse_accented_map
= {}
1159 for k
in accented_map
:
1160 inverse_accented_map
[accented_map
[k
]] = k
1162 # Since LyX may insert a line break within a word we must combine all
1163 # words before unicode normalization.
1164 # We do this only if the next line starts with an accent, otherwise we
1165 # would create things like '\begin_inset ERTstatus'.
1166 numberoflines
= len(document
.body
)
1167 for i
in range(numberoflines
-1):
1168 if document
.body
[i
] == '' or document
.body
[i
+1] == '' or document
.body
[i
][-1] == ' ':
1170 if (document
.body
[i
+1][0] in inverse_accent_map
):
1171 # the last character of this line and the first of the next line
1172 # form probably a surrogate pair.
1173 while (len(document
.body
[i
+1]) > 0 and document
.body
[i
+1][0] != ' '):
1174 document
.body
[i
] += document
.body
[i
+1][0]
1175 document
.body
[i
+1] = document
.body
[i
+1][1:]
1177 # Normalize to "Normal form D" (NFD, also known as canonical decomposition).
1178 # This is needed to catch all accented characters.
1179 for i
in range(numberoflines
):
1180 # Unfortunately we have a mixture of unicode strings and plain strings,
1181 # because we never use u'xxx' for string literals, but 'xxx'.
1182 # Therefore we may have to try two times to normalize the data.
1184 document
.body
[i
] = unicodedata
.normalize("NFD", document
.body
[i
])
1186 document
.body
[i
] = unicodedata
.normalize("NFD", unicode(document
.body
[i
], 'utf-8'))
1188 # Replace accented characters with InsetLaTeXAccent
1189 # Do not convert characters that can be represented in the chosen
1191 encoding_stack
= [get_encoding(document
.language
, document
.inputencoding
, 248, document
.cjk_encoding
)]
1192 lang_re
= re
.compile(r
"^\\lang\s(\S+)")
1194 while i
< len(document
.body
):
1196 if (document
.inputencoding
== "auto" or document
.inputencoding
== "default") and document
.cjk_encoding
!= '':
1197 # Track the encoding of the current line
1198 result
= lang_re
.match(document
.body
[i
])
1200 language
= result
.group(1)
1201 if language
== "default":
1202 encoding_stack
[-1] = document
.encoding
1204 from lyx2lyx_lang
import lang
1205 encoding_stack
[-1] = lang
[language
][3]
1207 elif find_token(document
.body
, "\\begin_layout", i
, i
+ 1) == i
:
1208 encoding_stack
.append(encoding_stack
[-1])
1210 elif find_token(document
.body
, "\\end_layout", i
, i
+ 1) == i
:
1211 del encoding_stack
[-1]
1214 for j
in range(len(document
.body
[i
])):
1215 # dotless i and dotless j are both in special_accent_map and can
1216 # occur as an accented character, so we need to test that the
1217 # following character is no accent
1218 if (document
.body
[i
][j
] in inverse_special_accent_map
and
1219 (j
== len(document
.body
[i
]) - 1 or document
.body
[i
][j
+1] not in inverse_accent_map
)):
1220 accent
= document
.body
[i
][j
]
1222 dummy
= accent
.encode(encoding_stack
[-1])
1223 except UnicodeEncodeError:
1224 # Insert the rest of the line as new line
1225 if j
< len(document
.body
[i
]) - 1:
1226 document
.body
.insert(i
+1, document
.body
[i
][j
+1:])
1227 # Delete the accented character
1229 document
.body
[i
] = document
.body
[i
][:j
-1]
1231 document
.body
[i
] = u
''
1232 # Finally add the InsetLaTeXAccent
1233 document
.body
[i
] += "\\i \\%s{}" % inverse_special_accent_map
[accent
]
1235 elif j
> 0 and document
.body
[i
][j
] in inverse_accent_map
:
1236 accented_char
= document
.body
[i
][j
-1]
1237 if accented_char
== ' ':
1238 # Conform to LyX output
1240 elif accented_char
in inverse_accented_map
:
1241 accented_char
= inverse_accented_map
[accented_char
]
1242 accent
= document
.body
[i
][j
]
1244 dummy
= unicodedata
.normalize("NFC", accented_char
+ accent
).encode(encoding_stack
[-1])
1245 except UnicodeEncodeError:
1246 # Insert the rest of the line as new line
1247 if j
< len(document
.body
[i
]) - 1:
1248 document
.body
.insert(i
+1, document
.body
[i
][j
+1:])
1249 # Delete the accented characters
1251 document
.body
[i
] = document
.body
[i
][:j
-2]
1253 document
.body
[i
] = u
''
1254 # Finally add the InsetLaTeXAccent
1255 document
.body
[i
] += "\\i \\%s{%s}" % (inverse_accent_map
[accent
], accented_char
)
1259 # Normalize to "Normal form C" (NFC, pre-composed characters) again
1260 for i
in range(numberoflines
):
1261 document
.body
[i
] = unicodedata
.normalize("NFC", document
.body
[i
])
1264 def normalize_font_whitespace_259(document
):
1265 """ Before format 259 the font changes were ignored if a
1266 whitespace was the first or last character in the sequence, this function
1267 transfers the whitespace outside."""
1269 char_properties
= {"\\series": "default",
1270 "\\emph": "default",
1272 "\\shape": "default",
1274 "\\family": "default"}
1275 return normalize_font_whitespace(document
, char_properties
)
1277 def normalize_font_whitespace_274(document
):
1278 """ Before format 259 (sic) the font changes were ignored if a
1279 whitespace was the first or last character in the sequence. This was
1280 corrected for most font properties in format 259, but the language
1281 was forgotten then. This function applies the same conversion done
1282 there (namely, transfers the whitespace outside) for font language
1283 changes, as well."""
1285 char_properties
= {"\\lang": "default"}
1286 return normalize_font_whitespace(document
, char_properties
)
1288 def get_paragraph_language(document
, i
):
1289 """ Return the language of the paragraph in which line i of the document
1290 body is. If the first thing in the paragraph is a \\lang command, that
1291 is the paragraph's langauge; otherwise, the paragraph's language is the
1292 document's language."""
1294 lines
= document
.body
1296 first_nonempty_line
= \
1297 find_nonempty_line(lines
, find_beginning_of_layout(lines
, i
) + 1)
1299 words
= lines
[first_nonempty_line
].split()
1301 if len(words
) > 1 and words
[0] == "\\lang":
1304 return document
.language
1306 def normalize_font_whitespace(document
, char_properties
):
1307 """ Before format 259 the font changes were ignored if a
1308 whitespace was the first or last character in the sequence, this function
1309 transfers the whitespace outside. Only a change in one of the properties
1310 in the provided char_properties is handled by this function."""
1312 if document
.backend
!= "latex":
1315 lines
= document
.body
1320 while i
< len(lines
):
1321 words
= lines
[i
].split()
1323 if len(words
) > 0 and words
[0] == "\\begin_layout":
1324 # a new paragraph resets all font changes
1326 # also reset the default language to be the paragraph's language
1327 if "\\lang" in char_properties
.keys():
1328 char_properties
["\\lang"] = \
1329 get_paragraph_language(document
, i
+ 1)
1331 elif len(words
) > 1 and words
[0] in char_properties
.keys():
1332 # we have a font change
1333 if char_properties
[words
[0]] == words
[1]:
1334 # property gets reset
1335 if words
[0] in changes
.keys():
1336 del changes
[words
[0]]
1337 defaultproperty
= True
1340 changes
[words
[0]] = words
[1]
1341 defaultproperty
= False
1343 # We need to explicitly reset all changed properties if we find
1344 # a space below, because LyX 1.4 would output the space after
1345 # closing the previous change and before starting the new one,
1346 # and closing a font change means to close all properties, not
1347 # just the changed one.
1349 if lines
[i
-1] and lines
[i
-1][-1] == " ":
1350 lines
[i
-1] = lines
[i
-1][:-1]
1351 # a space before the font change
1353 for k
in changes
.keys():
1354 # exclude property k because that is already in lines[i]
1356 added_lines
[1:1] = ["%s %s" % (k
, changes
[k
])]
1357 for k
in changes
.keys():
1358 # exclude property k because that must be added below anyway
1360 added_lines
[0:0] = ["%s %s" % (k
, char_properties
[k
])]
1362 # Property is reset in lines[i], so add the new stuff afterwards
1363 lines
[i
+1:i
+1] = added_lines
1365 # Reset property for the space
1366 added_lines
[0:0] = ["%s %s" % (words
[0], char_properties
[words
[0]])]
1367 lines
[i
:i
] = added_lines
1368 i
= i
+ len(added_lines
)
1370 elif lines
[i
+1] and lines
[i
+1][0] == " " and (len(changes
) > 0 or not defaultproperty
):
1371 # a space after the font change
1372 if (lines
[i
+1] == " " and lines
[i
+2]):
1373 next_words
= lines
[i
+2].split()
1374 if len(next_words
) > 0 and next_words
[0] == words
[0]:
1375 # a single blank with a property different from the
1376 # previous and the next line must not be changed
1379 lines
[i
+1] = lines
[i
+1][1:]
1381 for k
in changes
.keys():
1382 # exclude property k because that is already in lines[i]
1384 added_lines
[1:1] = ["%s %s" % (k
, changes
[k
])]
1385 for k
in changes
.keys():
1386 # exclude property k because that must be added below anyway
1388 added_lines
[0:0] = ["%s %s" % (k
, char_properties
[k
])]
1389 # Reset property for the space
1390 added_lines
[0:0] = ["%s %s" % (words
[0], char_properties
[words
[0]])]
1391 lines
[i
:i
] = added_lines
1392 i
= i
+ len(added_lines
)
1397 def revert_utf8x(document
):
1398 " Set utf8x encoding to utf8. "
1399 i
= find_token(document
.header
, "\\inputencoding", 0)
1401 document
.header
.append("\\inputencoding auto")
1403 inputenc
= get_value(document
.header
, "\\inputencoding", i
)
1404 if inputenc
== "utf8x":
1405 document
.header
[i
] = "\\inputencoding utf8"
1406 document
.inputencoding
= get_value(document
.header
, "\\inputencoding", 0)
1409 def revert_utf8plain(document
):
1410 " Set utf8plain encoding to utf8. "
1411 i
= find_token(document
.header
, "\\inputencoding", 0)
1413 document
.header
.append("\\inputencoding auto")
1415 inputenc
= get_value(document
.header
, "\\inputencoding", i
)
1416 if inputenc
== "utf8-plain":
1417 document
.header
[i
] = "\\inputencoding utf8"
1418 document
.inputencoding
= get_value(document
.header
, "\\inputencoding", 0)
1421 def revert_beamer_alert(document
):
1422 " Revert beamer's \\alert inset back to ERT. "
1425 i
= find_token(document
.body
, "\\begin_inset CharStyle Alert", i
)
1428 document
.body
[i
] = "\\begin_inset ERT"
1431 if (document
.body
[i
][:13] == "\\begin_layout"):
1432 # Insert the \alert command
1433 document
.body
[i
+ 1] = "\\alert{" + document
.body
[i
+ 1] + '}'
1440 def revert_beamer_structure(document
):
1441 " Revert beamer's \\structure inset back to ERT. "
1444 i
= find_token(document
.body
, "\\begin_inset CharStyle Structure", i
)
1447 document
.body
[i
] = "\\begin_inset ERT"
1450 if (document
.body
[i
][:13] == "\\begin_layout"):
1451 document
.body
[i
+ 1] = "\\structure{" + document
.body
[i
+ 1] + '}'
1458 def convert_changes(document
):
1459 " Switch output_changes off if tracking_changes is off. "
1460 i
= find_token(document
.header
, '\\tracking_changes', 0)
1462 document
.warning("Malformed lyx document: Missing '\\tracking_changes'.")
1464 j
= find_token(document
.header
, '\\output_changes', 0)
1466 document
.warning("Malformed lyx document: Missing '\\output_changes'.")
1468 tracking_changes
= get_value(document
.header
, "\\tracking_changes", i
)
1469 output_changes
= get_value(document
.header
, "\\output_changes", j
)
1470 if tracking_changes
== "false" and output_changes
== "true":
1471 document
.header
[j
] = "\\output_changes false"
1474 def revert_ascii(document
):
1475 " Set ascii encoding to auto. "
1476 i
= find_token(document
.header
, "\\inputencoding", 0)
1478 document
.header
.append("\\inputencoding auto")
1480 inputenc
= get_value(document
.header
, "\\inputencoding", i
)
1481 if inputenc
== "ascii":
1482 document
.header
[i
] = "\\inputencoding auto"
1483 document
.inputencoding
= get_value(document
.header
, "\\inputencoding", 0)
1486 def normalize_language_name(document
):
1487 lang
= { "brazil": "brazilian",
1488 "portuges": "portuguese"}
1490 if document
.language
in lang
:
1491 document
.language
= lang
[document
.language
]
1492 i
= find_token(document
.header
, "\\language", 0)
1493 document
.header
[i
] = "\\language %s" % document
.language
1496 def revert_language_name(document
):
1497 lang
= { "brazilian": "brazil",
1498 "portuguese": "portuges"}
1500 if document
.language
in lang
:
1501 document
.language
= lang
[document
.language
]
1502 i
= find_token(document
.header
, "\\language", 0)
1503 document
.header
[i
] = "\\language %s" % document
.language
1506 # \textclass cv -> \textclass simplecv
1507 def convert_cv_textclass(document
):
1508 if document
.textclass
== "cv":
1509 document
.textclass
= "simplecv"
1512 def revert_cv_textclass(document
):
1513 if document
.textclass
== "simplecv":
1514 document
.textclass
= "cv"
1518 # add scaleBeforeRotation graphics param
1519 def convert_graphics_rotation(document
):
1520 " add scaleBeforeRotation graphics parameter. "
1523 i
= find_token(document
.body
, "\\begin_inset Graphics", i
)
1526 j
= find_end_of_inset(document
.body
, i
+1)
1529 document
.warning("Malformed LyX document: Could not find end of graphics inset.")
1530 # Seach for rotateAngle and width or height or scale
1531 # If these params are not there, nothing needs to be done.
1532 k
= find_token(document
.body
, "\trotateAngle", i
+ 1, j
)
1533 l
= find_tokens(document
.body
, ["\twidth", "\theight", "\tscale"], i
+ 1, j
)
1534 if (k
!= -1 and l
!= -1):
1535 document
.body
.insert(j
, 'scaleBeforeRotation')
1540 # remove scaleBeforeRotation graphics param
1541 def revert_graphics_rotation(document
):
1542 " remove scaleBeforeRotation graphics parameter. "
1545 i
= find_token(document
.body
, "\\begin_inset Graphics", i
)
1548 j
= find_end_of_inset(document
.body
, i
+ 1)
1551 document
.warning("Malformed LyX document: Could not find end of graphics inset.")
1552 # If there's a scaleBeforeRotation param, just remove that
1553 k
= find_token(document
.body
, "\tscaleBeforeRotation", i
+ 1, j
)
1555 del document
.body
[k
]
1557 # if not, and if we have rotateAngle and width or height or scale,
1558 # we have to put the rotateAngle value to special
1559 rotateAngle
= get_value(document
.body
, 'rotateAngle', i
+ 1, j
)
1560 special
= get_value(document
.body
, 'special', i
+ 1, j
)
1561 if rotateAngle
!= "":
1562 k
= find_tokens(document
.body
, ["\twidth", "\theight", "\tscale"], i
+ 1, j
)
1566 document
.body
.insert(j
-1, '\tspecial angle=%s' % rotateAngle
)
1568 l
= find_token(document
.body
, "\tspecial", i
+ 1, j
)
1569 document
.body
[l
] = document
.body
[l
].replace(special
, 'angle=%s,%s' % (rotateAngle
, special
))
1570 k
= find_token(document
.body
, "\trotateAngle", i
+ 1, j
)
1572 del document
.body
[k
]
1577 def convert_tableborder(document
):
1578 # The problematic is: LyX double the table cell border as it ignores the "|" character in
1579 # the cell arguments. A fix takes care of this and therefore the "|" has to be removed
1581 while i
< len(document
.body
):
1582 h
= document
.body
[i
].find("leftline=\"true\"", 0, len(document
.body
[i
]))
1583 k
= document
.body
[i
].find("|>{", 0, len(document
.body
[i
]))
1584 # the two tokens have to be in one line
1585 if (h
!= -1 and k
!= -1):
1587 document
.body
[i
] = document
.body
[i
][:k
] + document
.body
[i
][k
+1:len(document
.body
[i
])-1]
1591 def revert_tableborder(document
):
1593 while i
< len(document
.body
):
1594 h
= document
.body
[i
].find("leftline=\"true\"", 0, len(document
.body
[i
]))
1595 k
= document
.body
[i
].find(">{", 0, len(document
.body
[i
]))
1596 # the two tokens have to be in one line
1597 if (h
!= -1 and k
!= -1):
1599 document
.body
[i
] = document
.body
[i
][:k
] + '|' + document
.body
[i
][k
:]
1603 def revert_armenian(document
):
1605 # set inputencoding from armscii8 to auto
1606 if document
.inputencoding
== "armscii8":
1607 i
= find_token(document
.header
, "\\inputencoding", 0)
1609 document
.header
[i
] = "\\inputencoding auto"
1610 # check if preamble exists, if not k is set to -1
1613 while i
< len(document
.preamble
):
1615 k
= document
.preamble
[i
].find("\\", 0, len(document
.preamble
[i
]))
1617 k
= document
.preamble
[i
].find("%", 0, len(document
.preamble
[i
]))
1619 # add the entry \usepackage{armtex} to the document preamble
1620 if document
.language
== "armenian":
1621 # set the armtex entry as the first preamble line
1623 document
.preamble
[0:0] = ["\\usepackage{armtex}"]
1624 # create the preamble when it doesn't exist
1626 document
.preamble
.append('\\usepackage{armtex}')
1627 # Set document language from armenian to english
1628 if document
.language
== "armenian":
1629 document
.language
= "english"
1630 i
= find_token(document
.header
, "\\language", 0)
1632 document
.header
[i
] = "\\language english"
1635 def revert_CJK(document
):
1636 " Set CJK encodings to default and languages chinese, japanese and korean to english. "
1637 encodings
= ["Bg5", "Bg5+", "GB", "GBt", "GBK", "JIS",
1638 "KS", "SJIS", "UTF8", "EUC-TW", "EUC-JP"]
1639 i
= find_token(document
.header
, "\\inputencoding", 0)
1641 document
.header
.append("\\inputencoding auto")
1643 inputenc
= get_value(document
.header
, "\\inputencoding", i
)
1644 if inputenc
in encodings
:
1645 document
.header
[i
] = "\\inputencoding default"
1646 document
.inputencoding
= get_value(document
.header
, "\\inputencoding", 0)
1648 if document
.language
== "chinese-simplified" or \
1649 document
.language
== "chinese-traditional" or \
1650 document
.language
== "japanese" or document
.language
== "korean":
1651 document
.language
= "english"
1652 i
= find_token(document
.header
, "\\language", 0)
1654 document
.header
[i
] = "\\language english"
1657 def revert_preamble_listings_params(document
):
1658 " Revert preamble option \listings_params "
1659 i
= find_token(document
.header
, "\\listings_params", 0)
1661 document
.preamble
.append('\\usepackage{listings}')
1662 document
.preamble
.append('\\lstset{%s}' % document
.header
[i
].split()[1].strip('"'))
1663 document
.header
.pop(i
);
1666 def revert_listings_inset(document
):
1667 r
''' Revert listings inset to \lstinline or \begin, \end lstlisting, translate
1671 lstparams "language=Delphi"
1675 \begin_layout Standard
1685 \begin_layout Standard
1689 lstinline[language=Delphi]{var i = 10;}
1694 There can be an caption inset in this inset
1696 \begin_layout Standard
1697 \begin_inset Caption
1699 \begin_layout Standard
1701 \begin_inset LatexCommand label
1717 i
= find_token(document
.body
, '\\begin_inset listings', i
)
1721 if not '\\usepackage{listings}' in document
.preamble
:
1722 document
.preamble
.append('\\usepackage{listings}')
1723 j
= find_end_of_inset(document
.body
, i
+ 1)
1725 # this should not happen
1731 for line
in range(i
+ 1, i
+ 4):
1732 if document
.body
[line
].startswith('inline'):
1733 inline
= document
.body
[line
].split()[1]
1734 if document
.body
[line
].startswith('lstparams'):
1735 params
= document
.body
[line
].split()[1].strip('"')
1736 if document
.body
[line
].startswith('status'):
1737 status
= document
.body
[line
].split()[1].strip()
1742 cap
= find_token(document
.body
, '\\begin_inset Caption', i
)
1744 cap_end
= find_end_of_inset(document
.body
, cap
+ 1)
1746 # this should not happen
1749 lbl
= find_token(document
.body
, '\\begin_inset LatexCommand label', cap
+ 1)
1751 lbl_end
= find_end_of_inset(document
.body
, lbl
+ 1)
1753 # this should not happen
1758 for line
in document
.body
[lbl
: lbl_end
+ 1]:
1759 if line
.startswith('name '):
1760 label
= line
.split()[1].strip('"')
1762 for line
in document
.body
[cap
: lbl
] + document
.body
[lbl_end
+ 1 : cap_end
+ 1]:
1763 if not line
.startswith('\\'):
1764 caption
+= line
.strip()
1767 # looking for the oneline code for lstinline
1768 inlinecode
= document
.body
[find_end_of_layout(document
.body
,
1769 find_token(document
.body
, '\\begin_layout %s' % document
.default_layout
, i
+ 1) +1 ) - 1]
1770 if len(caption
) > 0:
1771 if len(params
) == 0:
1772 params
= 'caption={%s}' % caption
1774 params
+= ',caption={%s}' % caption
1776 if len(params
) == 0:
1777 params
= 'label={%s}' % label
1779 params
+= ',label={%s}' % label
1781 params
= '[%s]' % params
1782 params
= params
.replace('\\', '\\backslash\n')
1783 if inline
== 'true':
1784 document
.body
[i
:(j
+1)] = [r
'\begin_inset ERT',
1785 'status %s' % status
,
1786 r
'\begin_layout %s' % document
.default_layout
,
1790 'lstinline%s{%s}' % (params
, inlinecode
),
1795 document
.body
[i
: j
+1] = [r
'\begin_inset ERT',
1796 'status %s' % status
,
1798 r
'\begin_layout %s' % document
.default_layout
,
1802 r
'begin{lstlisting}%s' % params
,
1804 ] + document
.body
[k
: j
- 1] + \
1806 r
'\begin_layout %s' % document
.default_layout
,
1815 def revert_include_listings(document
):
1816 r
''' Revert lstinputlisting Include option , translate
1817 \begin_inset Include \lstinputlisting{file}[opt]
1827 \begin_layout Standard
1831 lstinputlisting{file}[opt]
1839 i
= find_token(document
.body
, r
'\begin_inset Include \lstinputlisting', i
)
1843 if not '\\usepackage{listings}' in document
.preamble
:
1844 document
.preamble
.append('\\usepackage{listings}')
1845 j
= find_end_of_inset(document
.body
, i
+ 1)
1847 # this should not happen
1849 # find command line lstinputlisting{file}[options]
1850 cmd
, file, option
= '', '', ''
1851 if re
.match(r
'\\(lstinputlisting){([.\w]*)}(.*)', document
.body
[i
].split()[2]):
1852 cmd
, file, option
= re
.match(r
'\\(lstinputlisting){([.\w]*)}(.*)', document
.body
[i
].split()[2]).groups()
1853 option
= option
.replace('\\', '\\backslash\n')
1854 document
.body
[i
: j
+ 1] = [r
'\begin_inset ERT',
1857 r
'\begin_layout %s' % document
.default_layout
,
1861 '%s%s{%s}' % (cmd
, option
, file),
1867 def revert_ext_font_sizes(document
):
1868 if document
.backend
!= "latex": return
1869 if not document
.textclass
.startswith("ext"): return
1871 fontsize
= get_value(document
.header
, '\\paperfontsize', 0)
1872 if fontsize
not in ('10', '11', '12'): return
1875 i
= find_token(document
.header
, '\\paperfontsize', 0)
1876 document
.header
[i
] = '\\paperfontsize default'
1878 i
= find_token(document
.header
, '\\options', 0)
1880 i
= find_token(document
.header
, '\\textclass', 0) + 1
1881 document
.header
[i
:i
] = ['\\options %s' % fontsize
]
1883 document
.header
[i
] += ',%s' % fontsize
1886 def convert_ext_font_sizes(document
):
1887 if document
.backend
!= "latex": return
1888 if not document
.textclass
.startswith("ext"): return
1890 fontsize
= get_value(document
.header
, '\\paperfontsize', 0)
1891 if fontsize
!= 'default': return
1893 i
= find_token(document
.header
, '\\options', 0)
1896 options
= get_value(document
.header
, '\\options', i
)
1898 fontsizes
= '10pt', '11pt', '12pt'
1899 for fs
in fontsizes
:
1900 if options
.find(fs
) != -1:
1902 else: # this else will only be attained if the for cycle had no match
1905 options
= options
.split(',')
1906 for j
, opt
in enumerate(options
):
1907 if opt
in fontsizes
:
1914 k
= find_token(document
.header
, '\\paperfontsize', 0)
1915 document
.header
[k
] = '\\paperfontsize %s' % fontsize
1918 document
.header
[i
] = '\\options %s' % ','.join(options
)
1920 del document
.header
[i
]
1923 def revert_separator_layout(document
):
1924 r
'''Revert --Separator-- to a lyx note
1927 \begin_layout --Separator--
1933 \begin_layout Standard
1934 \begin_inset Note Note
1937 \begin_layout Standard
1950 i
= find_token(document
.body
, r
'\begin_layout --Separator--', i
)
1953 j
= find_end_of_layout(document
.body
, i
+ 1)
1955 # this should not happen
1957 document
.body
[i
: j
+ 1] = [r
'\begin_layout %s' % document
.default_layout
,
1958 r
'\begin_inset Note Note',
1961 r
'\begin_layout %s' % document
.default_layout
,
1962 'Separate Environment',
1966 document
.body
[ i
+ 1 : j
] + \
1972 def convert_arabic (document
):
1973 if document
.language
== "arabic":
1974 document
.language
= "arabic_arabtex"
1975 i
= find_token(document
.header
, "\\language", 0)
1977 document
.header
[i
] = "\\language arabic_arabtex"
1979 while i
< len(document
.body
):
1980 h
= document
.body
[i
].find("\lang arabic", 0, len(document
.body
[i
]))
1982 # change the language name
1983 document
.body
[i
] = '\lang arabic_arabtex'
1987 def revert_arabic (document
):
1988 if document
.language
== "arabic_arabtex":
1989 document
.language
= "arabic"
1990 i
= find_token(document
.header
, "\\language", 0)
1992 document
.header
[i
] = "\\language arabic"
1994 while i
< len(document
.body
):
1995 h
= document
.body
[i
].find("\lang arabic_arabtex", 0, len(document
.body
[i
]))
1997 # change the language name
1998 document
.body
[i
] = '\lang arabic'
2006 supported_versions
= ["1.5.0","1.5"]
2007 convert
= [[246, []],
2008 [247, [convert_font_settings
]],
2010 [249, [convert_utf8
]],
2013 [252, [convert_commandparams
, convert_bibitem
]],
2015 [254, [convert_esint
]],
2018 [257, [convert_caption
]],
2019 [258, [convert_lyxline
]],
2020 [259, [convert_accent
, normalize_font_whitespace_259
]],
2022 [261, [convert_changes
]],
2024 [263, [normalize_language_name
]],
2025 [264, [convert_cv_textclass
]],
2026 [265, [convert_tableborder
]],
2032 [271, [convert_ext_font_sizes
]],
2035 [274, [normalize_font_whitespace_274
]],
2036 [275, [convert_graphics_rotation
]],
2037 [276, [convert_arabic
]]
2041 [275, [revert_arabic
]],
2042 [274, [revert_graphics_rotation
]],
2044 [272, [revert_separator_layout
]],
2045 [271, [revert_preamble_listings_params
, revert_listings_inset
, revert_include_listings
]],
2046 [270, [revert_ext_font_sizes
]],
2047 [269, [revert_beamer_alert
, revert_beamer_structure
]],
2048 [268, [revert_preamble_listings_params
, revert_listings_inset
, revert_include_listings
]],
2049 [267, [revert_CJK
]],
2050 [266, [revert_utf8plain
]],
2051 [265, [revert_armenian
]],
2052 [264, [revert_tableborder
]],
2053 [263, [revert_cv_textclass
]],
2054 [262, [revert_language_name
]],
2055 [261, [revert_ascii
]],
2057 [259, [revert_utf8x
]],
2060 [256, [revert_caption
]],
2061 [255, [revert_encodings
]],
2062 [254, [revert_clearpage
, revert_cleardoublepage
]],
2063 [253, [revert_esint
]],
2064 [252, [revert_nomenclature
, revert_printnomenclature
]],
2065 [251, [revert_commandparams
]],
2066 [250, [revert_cs_label
]],
2068 [248, [revert_accent
, revert_utf8
, revert_unicode
]],
2069 [247, [revert_booktabs
]],
2070 [246, [revert_font_settings
]],
2071 [245, [revert_framed
]]]
2074 if __name__
== "__main__":