|
1 import re |
|
2 from django.conf import settings |
|
3 from django.utils.encoding import force_unicode |
|
4 from django.utils.functional import allow_lazy |
|
5 from django.utils.translation import ugettext_lazy |
|
6 from htmlentitydefs import name2codepoint |
|
7 |
|
8 # Capitalizes the first letter of a string. |
|
9 capfirst = lambda x: x and force_unicode(x)[0].upper() + force_unicode(x)[1:] |
|
10 capfirst = allow_lazy(capfirst, unicode) |
|
11 |
|
12 def wrap(text, width): |
|
13 """ |
|
14 A word-wrap function that preserves existing line breaks and most spaces in |
|
15 the text. Expects that existing line breaks are posix newlines. |
|
16 """ |
|
17 text = force_unicode(text) |
|
18 def _generator(): |
|
19 it = iter(text.split(' ')) |
|
20 word = it.next() |
|
21 yield word |
|
22 pos = len(word) - word.rfind('\n') - 1 |
|
23 for word in it: |
|
24 if "\n" in word: |
|
25 lines = word.split('\n') |
|
26 else: |
|
27 lines = (word,) |
|
28 pos += len(lines[0]) + 1 |
|
29 if pos > width: |
|
30 yield '\n' |
|
31 pos = len(lines[-1]) |
|
32 else: |
|
33 yield ' ' |
|
34 if len(lines) > 1: |
|
35 pos = len(lines[-1]) |
|
36 yield word |
|
37 return u''.join(_generator()) |
|
38 wrap = allow_lazy(wrap, unicode) |
|
39 |
|
40 def truncate_words(s, num): |
|
41 "Truncates a string after a certain number of words." |
|
42 s = force_unicode(s) |
|
43 length = int(num) |
|
44 words = s.split() |
|
45 if len(words) > length: |
|
46 words = words[:length] |
|
47 if not words[-1].endswith('...'): |
|
48 words.append('...') |
|
49 return u' '.join(words) |
|
50 truncate_words = allow_lazy(truncate_words, unicode) |
|
51 |
|
52 def truncate_html_words(s, num): |
|
53 """ |
|
54 Truncates html to a certain number of words (not counting tags and |
|
55 comments). Closes opened tags if they were correctly closed in the given |
|
56 html. |
|
57 """ |
|
58 s = force_unicode(s) |
|
59 length = int(num) |
|
60 if length <= 0: |
|
61 return u'' |
|
62 html4_singlets = ('br', 'col', 'link', 'base', 'img', 'param', 'area', 'hr', 'input') |
|
63 # Set up regular expressions |
|
64 re_words = re.compile(r'&.*?;|<.*?>|(\w[\w-]*)', re.U) |
|
65 re_tag = re.compile(r'<(/)?([^ ]+?)(?: (/)| .*?)?>') |
|
66 # Count non-HTML words and keep note of open tags |
|
67 pos = 0 |
|
68 ellipsis_pos = 0 |
|
69 words = 0 |
|
70 open_tags = [] |
|
71 while words <= length: |
|
72 m = re_words.search(s, pos) |
|
73 if not m: |
|
74 # Checked through whole string |
|
75 break |
|
76 pos = m.end(0) |
|
77 if m.group(1): |
|
78 # It's an actual non-HTML word |
|
79 words += 1 |
|
80 if words == length: |
|
81 ellipsis_pos = pos |
|
82 continue |
|
83 # Check for tag |
|
84 tag = re_tag.match(m.group(0)) |
|
85 if not tag or ellipsis_pos: |
|
86 # Don't worry about non tags or tags after our truncate point |
|
87 continue |
|
88 closing_tag, tagname, self_closing = tag.groups() |
|
89 tagname = tagname.lower() # Element names are always case-insensitive |
|
90 if self_closing or tagname in html4_singlets: |
|
91 pass |
|
92 elif closing_tag: |
|
93 # Check for match in open tags list |
|
94 try: |
|
95 i = open_tags.index(tagname) |
|
96 except ValueError: |
|
97 pass |
|
98 else: |
|
99 # SGML: An end tag closes, back to the matching start tag, all unclosed intervening start tags with omitted end tags |
|
100 open_tags = open_tags[i+1:] |
|
101 else: |
|
102 # Add it to the start of the open tags list |
|
103 open_tags.insert(0, tagname) |
|
104 if words <= length: |
|
105 # Don't try to close tags if we don't need to truncate |
|
106 return s |
|
107 out = s[:ellipsis_pos] + ' ...' |
|
108 # Close any tags still open |
|
109 for tag in open_tags: |
|
110 out += '</%s>' % tag |
|
111 # Return string |
|
112 return out |
|
113 truncate_html_words = allow_lazy(truncate_html_words, unicode) |
|
114 |
|
115 def get_valid_filename(s): |
|
116 """ |
|
117 Returns the given string converted to a string that can be used for a clean |
|
118 filename. Specifically, leading and trailing spaces are removed; other |
|
119 spaces are converted to underscores; and anything that is not a unicode |
|
120 alphanumeric, dash, underscore, or dot, is removed. |
|
121 >>> get_valid_filename("john's portrait in 2004.jpg") |
|
122 u'johns_portrait_in_2004.jpg' |
|
123 """ |
|
124 s = force_unicode(s).strip().replace(' ', '_') |
|
125 return re.sub(r'(?u)[^-\w.]', '', s) |
|
126 get_valid_filename = allow_lazy(get_valid_filename, unicode) |
|
127 |
|
128 def get_text_list(list_, last_word=ugettext_lazy(u'or')): |
|
129 """ |
|
130 >>> get_text_list(['a', 'b', 'c', 'd']) |
|
131 u'a, b, c or d' |
|
132 >>> get_text_list(['a', 'b', 'c'], 'and') |
|
133 u'a, b and c' |
|
134 >>> get_text_list(['a', 'b'], 'and') |
|
135 u'a and b' |
|
136 >>> get_text_list(['a']) |
|
137 u'a' |
|
138 >>> get_text_list([]) |
|
139 u'' |
|
140 """ |
|
141 if len(list_) == 0: return u'' |
|
142 if len(list_) == 1: return force_unicode(list_[0]) |
|
143 return u'%s %s %s' % (', '.join([force_unicode(i) for i in list_][:-1]), force_unicode(last_word), force_unicode(list_[-1])) |
|
144 get_text_list = allow_lazy(get_text_list, unicode) |
|
145 |
|
146 def normalize_newlines(text): |
|
147 return force_unicode(re.sub(r'\r\n|\r|\n', '\n', text)) |
|
148 normalize_newlines = allow_lazy(normalize_newlines, unicode) |
|
149 |
|
150 def recapitalize(text): |
|
151 "Recapitalizes text, placing caps after end-of-sentence punctuation." |
|
152 text = force_unicode(text).lower() |
|
153 capsRE = re.compile(r'(?:^|(?<=[\.\?\!] ))([a-z])') |
|
154 text = capsRE.sub(lambda x: x.group(1).upper(), text) |
|
155 return text |
|
156 recapitalize = allow_lazy(recapitalize) |
|
157 |
|
158 def phone2numeric(phone): |
|
159 "Converts a phone number with letters into its numeric equivalent." |
|
160 letters = re.compile(r'[A-PR-Y]', re.I) |
|
161 char2number = lambda m: {'a': '2', 'c': '2', 'b': '2', 'e': '3', |
|
162 'd': '3', 'g': '4', 'f': '3', 'i': '4', 'h': '4', 'k': '5', |
|
163 'j': '5', 'm': '6', 'l': '5', 'o': '6', 'n': '6', 'p': '7', |
|
164 's': '7', 'r': '7', 'u': '8', 't': '8', 'w': '9', 'v': '8', |
|
165 'y': '9', 'x': '9'}.get(m.group(0).lower()) |
|
166 return letters.sub(char2number, phone) |
|
167 phone2numeric = allow_lazy(phone2numeric) |
|
168 |
|
169 # From http://www.xhaus.com/alan/python/httpcomp.html#gzip |
|
170 # Used with permission. |
|
171 def compress_string(s): |
|
172 import cStringIO, gzip |
|
173 zbuf = cStringIO.StringIO() |
|
174 zfile = gzip.GzipFile(mode='wb', compresslevel=6, fileobj=zbuf) |
|
175 zfile.write(s) |
|
176 zfile.close() |
|
177 return zbuf.getvalue() |
|
178 |
|
179 ustring_re = re.compile(u"([\u0080-\uffff])") |
|
180 |
|
181 def javascript_quote(s, quote_double_quotes=False): |
|
182 |
|
183 def fix(match): |
|
184 return r"\u%04x" % ord(match.group(1)) |
|
185 |
|
186 if type(s) == str: |
|
187 s = s.decode('utf-8') |
|
188 elif type(s) != unicode: |
|
189 raise TypeError, s |
|
190 s = s.replace('\\', '\\\\') |
|
191 s = s.replace('\r', '\\r') |
|
192 s = s.replace('\n', '\\n') |
|
193 s = s.replace('\t', '\\t') |
|
194 s = s.replace("'", "\\'") |
|
195 if quote_double_quotes: |
|
196 s = s.replace('"', '"') |
|
197 return str(ustring_re.sub(fix, s)) |
|
198 javascript_quote = allow_lazy(javascript_quote, unicode) |
|
199 |
|
200 # Expression to match some_token and some_token="with spaces" (and similarly |
|
201 # for single-quoted strings). |
|
202 smart_split_re = re.compile(r""" |
|
203 ([^\s"]*"(?:[^"\\]*(?:\\.[^"\\]*)*)"\S*| |
|
204 [^\s']*'(?:[^'\\]*(?:\\.[^'\\]*)*)'\S*| |
|
205 \S+)""", re.VERBOSE) |
|
206 |
|
207 def smart_split(text): |
|
208 r""" |
|
209 Generator that splits a string by spaces, leaving quoted phrases together. |
|
210 Supports both single and double quotes, and supports escaping quotes with |
|
211 backslashes. In the output, strings will keep their initial and trailing |
|
212 quote marks and escaped quotes will remain escaped (the results can then |
|
213 be further processed with unescape_string_literal()). |
|
214 |
|
215 >>> list(smart_split(r'This is "a person\'s" test.')) |
|
216 [u'This', u'is', u'"a person\\\'s"', u'test.'] |
|
217 >>> list(smart_split(r"Another 'person\'s' test.")) |
|
218 [u'Another', u"'person\\'s'", u'test.'] |
|
219 >>> list(smart_split(r'A "\"funky\" style" test.')) |
|
220 [u'A', u'"\\"funky\\" style"', u'test.'] |
|
221 """ |
|
222 text = force_unicode(text) |
|
223 for bit in smart_split_re.finditer(text): |
|
224 yield bit.group(0) |
|
225 smart_split = allow_lazy(smart_split, unicode) |
|
226 |
|
227 def _replace_entity(match): |
|
228 text = match.group(1) |
|
229 if text[0] == u'#': |
|
230 text = text[1:] |
|
231 try: |
|
232 if text[0] in u'xX': |
|
233 c = int(text[1:], 16) |
|
234 else: |
|
235 c = int(text) |
|
236 return unichr(c) |
|
237 except ValueError: |
|
238 return match.group(0) |
|
239 else: |
|
240 try: |
|
241 return unichr(name2codepoint[text]) |
|
242 except (ValueError, KeyError): |
|
243 return match.group(0) |
|
244 |
|
245 _entity_re = re.compile(r"&(#?[xX]?(?:[0-9a-fA-F]+|\w{1,8}));") |
|
246 |
|
247 def unescape_entities(text): |
|
248 return _entity_re.sub(_replace_entity, text) |
|
249 unescape_entities = allow_lazy(unescape_entities, unicode) |
|
250 |
|
251 def unescape_string_literal(s): |
|
252 r""" |
|
253 Convert quoted string literals to unquoted strings with escaped quotes and |
|
254 backslashes unquoted:: |
|
255 |
|
256 >>> unescape_string_literal('"abc"') |
|
257 'abc' |
|
258 >>> unescape_string_literal("'abc'") |
|
259 'abc' |
|
260 >>> unescape_string_literal('"a \"bc\""') |
|
261 'a "bc"' |
|
262 >>> unescape_string_literal("'\'ab\' c'") |
|
263 "'ab' c" |
|
264 """ |
|
265 if s[0] not in "\"'" or s[-1] != s[0]: |
|
266 raise ValueError("Not a string literal: %r" % s) |
|
267 quote = s[0] |
|
268 return s[1:-1].replace(r'\%s' % quote, quote).replace(r'\\', '\\') |
|
269 unescape_string_literal = allow_lazy(unescape_string_literal) |