--- a/src/jocondelab/views/front_office.py Mon Dec 02 14:58:51 2013 +0100
+++ b/src/jocondelab/views/front_office.py Mon Dec 02 16:08:58 2013 +0100
@@ -118,7 +118,7 @@
if i == 0 and page == 1 and len(dbpedia_uris) == 1:
context["wkinfo"] = firstres
fs = fs.values('term_id').distinct()
- qs = qs.filter(noticeterm__term__in=fs)
+ qs = qs.filter(noticeterm__term__in=fs, noticeterm__term__validated=True)
if thesaurus:
if thesaurus == 'REPR':
qs = qs.filter(noticeterm__term__thesaurus__label__in=['REPR','SREP'])
@@ -137,11 +137,11 @@
if i == 0 and page == 1 and len(queryterms) == 1:
context["wkinfo"] = firstres
fs = fs.values('term_id').distinct()
- qs = qs.filter(noticeterm__term__in=fs)
+ qs = qs.filter(noticeterm__term__in=fs, noticeterm__term__validated=True)
elif from_year:
queryobj = {'from_year': from_year, 'to_year': to_year}
searchterms = [u"%s – %s"%(from_year, to_year)]
- qs = qs.filter(years__start_year__lte=to_year, years__end_year__gte=from_year)
+ qs = qs.filter(years__start_year__lte=to_year, years__end_year__gte=from_year, noticeterm__term__validated=True)
context["queryobj"] = json.dumps(queryobj)
context["querystr"] = urlencode(queryobj)
context["searchterms_label"] = ugettext(u" ET ").join(searchterms)
@@ -153,6 +153,8 @@
ids = paginator.page(page)
context["page_count"] = paginator.num_pages
context["count"] = paginator.count
+ if paginator.count==0:
+ context["wkinfo"] = None
# Now that we have the list of ids
ns = Notice.objects.filter(pk__in=ids).extra(select={'relative_url': '"core_noticeimage"."relative_url"'}).filter(image=True).filter(images__main=True).order_by('id')