# HG changeset patch # User cavaliet # Date 1385996938 -3600 # Node ID e684e2b830a9011c02c733af4425814415cfe214 # Parent 1797ef9ea6bf90f4d1e8828f822844fbeab9887e bug 0018227 diff -r 1797ef9ea6bf -r e684e2b830a9 src/jocondelab/views/front_office.py --- a/src/jocondelab/views/front_office.py Mon Dec 02 14:58:51 2013 +0100 +++ b/src/jocondelab/views/front_office.py Mon Dec 02 16:08:58 2013 +0100 @@ -118,7 +118,7 @@ if i == 0 and page == 1 and len(dbpedia_uris) == 1: context["wkinfo"] = firstres fs = fs.values('term_id').distinct() - qs = qs.filter(noticeterm__term__in=fs) + qs = qs.filter(noticeterm__term__in=fs, noticeterm__term__validated=True) if thesaurus: if thesaurus == 'REPR': qs = qs.filter(noticeterm__term__thesaurus__label__in=['REPR','SREP']) @@ -137,11 +137,11 @@ if i == 0 and page == 1 and len(queryterms) == 1: context["wkinfo"] = firstres fs = fs.values('term_id').distinct() - qs = qs.filter(noticeterm__term__in=fs) + qs = qs.filter(noticeterm__term__in=fs, noticeterm__term__validated=True) elif from_year: queryobj = {'from_year': from_year, 'to_year': to_year} searchterms = [u"%s – %s"%(from_year, to_year)] - qs = qs.filter(years__start_year__lte=to_year, years__end_year__gte=from_year) + qs = qs.filter(years__start_year__lte=to_year, years__end_year__gte=from_year, noticeterm__term__validated=True) context["queryobj"] = json.dumps(queryobj) context["querystr"] = urlencode(queryobj) context["searchterms_label"] = ugettext(u" ET ").join(searchterms) @@ -153,6 +153,8 @@ ids = paginator.page(page) context["page_count"] = paginator.num_pages context["count"] = paginator.count + if paginator.count==0: + context["wkinfo"] = None # Now that we have the list of ids ns = Notice.objects.filter(pk__in=ids).extra(select={'relative_url': '"core_noticeimage"."relative_url"'}).filter(image=True).filter(images__main=True).order_by('id')