Hostname: page-component-586b7cd67f-rcrh6 Total loading time: 0 Render date: 2024-11-23T20:20:35.508Z Has data issue: false hasContentIssue false

A Computer Vision Approach to Identify Einstein Rings and Arcs

Published online by Cambridge University Press:  31 March 2017

Chien-Hsiu Lee*
Affiliation:
Subaru Telescope, National Astronomical Observatory of Japan, 650 North Aohoku Place, Hilo, HI 96720, USA
Rights & Permissions [Opens in a new window]

Abstract

Einstein rings are rare gems of strong lensing phenomena; the ring images can be used to probe the underlying lens gravitational potential at every position angles, tightly constraining the lens mass profile. In addition, the magnified images also enable us to probe high-z galaxies with enhanced resolution and signal-to-noise ratios. However, only a handful of Einstein rings have been reported, either from serendipitous discoveries or or visual inspections of hundred thousands of massive galaxies or galaxy clusters. In the era of large sky surveys, an automated approach to identify ring pattern in the big data to come is in high demand. Here, we present an Einstein ring recognition approach based on computer vision techniques. The workhorse is the circle Hough transform that recognise circular patterns or arcs in the images. We propose a two-tier approach by first pre-selecting massive galaxies associated with multiple blue objects as possible lens, than use Hough transform to identify circular pattern. As a proof-of-concept, we apply our approach to SDSS, with a high completeness, albeit with low purity. We also apply our approach to other lenses in DES, HSC-SSP, and UltraVISTA survey, illustrating the versatility of our approach.

Type
Research Article
Copyright
Copyright © Astronomical Society of Australia 2017 

1 INTRODUCTION

According to Einstein’s theory of general relativity (Einstein Reference Einstein1915), massive foreground galaxies or clusters of galaxies induce strong space–time curvature, serving as gravitational lenses and bend the lights of background sources to form multiple and magnified images along the observer’s line-of-sight. Such strong lensing systems provide unique opportunities to study distant galaxies and cosmology in detail (see Treu Reference Treu2010, for a recent review). On the one hand, modelling the lensed images enables us to constrain the mass profile and dark matter distribution of the lens, as well as probing dwarf satellites associated with the lens galaxy, to test ΛCDM model (Vegetti et al. Reference Vegetti, Lagattuta, McKean, Auger, Fassnacht and Koopmans2012). On the other hand, gravitational lensing in tandem with Hubble space telescope offers us a peek into galaxies at cosmic dawn (Zheng et al. Reference Zheng2012), beyond the reach of current instruments. In addition, joining the forces of supernova and gravitational lensing will provide insights to the expansion rate of our Universe (Refsdal Reference Refsdal1964; Kelly et al. Reference Kelly2015).

When the foreground lens galaxy, the background source, and the observer are well aligned on a straight line, the lensed images will become a ring. Such phenomena were first pointed out by Chwolson (Reference Chwolson1924) and later remarked by Einstein (Reference Einstein1936) upon the request of R. W. Mandl, hence bear the name Einstein–Chwolson ring (hereafter Einstein ring, for simplicity). Einstein rings are rare gems of the strong lensing phenomena; they provide tighter constraints on the lens than multiply-imaged strong lensing systems, because (1) the enclosed lens mass can be directly measured from the angular size of the Einstein ring, without any prior assumptions on the lens mass profile and underlying lensing potential (see e.g. Narayan & Bartelmann Reference Narayan and Bartelmann1996); (2) the lens modelling, in tandem with lens velocity dispersion gradient derived from spatially resolved spectra of integral field unit spectrograph observations, will provide a complete view of the mass distribution of baryonic and dark matter; (3) compared to doubly or quadruply lensed systems, the gravitational potential of the lens can be studied in great detail, because the ring structure can provide constraints at virtually every position angle (Kochanek, Keeton, & McLeod Reference Kochanek, Keeton and McLeod2001); (4) the stretched and magnified source images offer us a unique opportunity to resolve the source with enhanced spatial resolution and higher signal-to-noise ratio, enabling us to study metallicity gradient across the source in detail (Jones et al. Reference Jones, Ellis, Richard and Jullo2013).

Despite their usefulness, only dozens of partial Einstein ring systems have been discovered so far (see e.g. Bolton et al. Reference Bolton, Burles, Koopmans, Treu, Gavazzi, Moustakas, Wayth and Schlegel2008a; Stark et al. Reference Stark2013), mainly due to the rare chance to have lenses and sources well-aligned along the observer’s line-of-sights. Among these partial Einstein ring systems, only a handful of them show complete or nearly complete ring morphology, e.g. B1938+666 (King et al. Reference King1998), FOR J0332-2557 (~ 260° Cabanac et al. Reference Cabanac, Valls-Gabaud, Jaunsen, Lidman and Jerjen2005), Cosmic Horseshoe (~ 300° Belokurov et al. Reference Belokurov2007), Elliot Arc (Buckley-Geer et al. Reference Buckley-Geer2011), Canarias Einstein ring (Bettinelli et al. Reference Bettinelli, Simioni, Aparicio, Hidalgo, Cassisi, Walker, Piotto and Valdes2016), Eye of Horus (Tanaka et al. Reference Tanaka2016), to name a few. These Einstein rings are detected either serendipitously (e.g. Cabanac et al. Reference Cabanac, Valls-Gabaud, Jaunsen, Lidman and Jerjen2005; Buckley-Geer et al. Reference Buckley-Geer2011; Bettinelli et al. Reference Bettinelli, Simioni, Aparicio, Hidalgo, Cassisi, Walker, Piotto and Valdes2016), or by examining images of massive galaxies as possible lens, containing tens of thousands of luminous red galaxies (LRGs) and brightest cluster galaxies (BCG) (e.g. Lin et al. Reference Lin2009; Tanaka et al. Reference Tanaka2016). With the advent of ultra-wide cameras and large area surveys, e.g. Sloan Digital Sky SurveyFootnote 1 (SDSS, York et al. Reference York2000), Pan-STARRSFootnote 2 (Kaiser et al. Reference Kaiser2010), SkyMapperFootnote 3 (Schmidt et al. Reference Schmidt, Keller, Francis and Bessell2005), Dark Energy SurveyFootnote 4 (DES, Flaugher et al. Reference Flaugher2015), and Hyper Suprime-Cam Subaru Strategic ProgramFootnote 5 (HSC, Miyazaki et al. Reference Miyazaki2006), and Large Synoptic Survey TelescopeFootnote 6 (LSST, LSST Science Collaboration et al. 2009), hundred thousands of lens are expected to be discovered (Collett Reference Collett2015); it is very difficult, if not impossible, to visually inspect all the candidate LRG and BCG lenses in the big data to comeFootnote 7 .

In a more general context, there are several attempts to identify gravitational arcs or multiply lensed objects around massive galaxies, using either image-based or spectroscopy-based manners. For example, Belokurov et al. (Reference Belokurov2007) have shown that by simply selecting bright, red galaxies associated with multiple blue, faint objects in SDSS already yields several strong lens candidates with wide arcs. They further tailored the selecting criteria for SDSS, in order to eliminate false-positives from galaxy groups, mergers, and tidal tails, and presented more promising arc candidates in Belokurov et al. (Reference Belokurov, Evans, Hewett, Moiseev, McMahon, Sanchez and King2009). More sophisticated image-based approach involves high-resolution images with dedicated modelling. For example, Marshall et al. (Reference Marshall, Hogg, Moustakas, Fassnacht, Bradač, Schrabback and Blandford2009), Gavazzi et al. (Reference Gavazzi, Marshall, Treu and Sonnenfeld2014), Brault & Gavazzi (Reference Brault and Gavazzi2015), and Paraficz et al. (Reference Paraficz2016) have demonstrated the feasibility to remove the foreground lens light with dedicated modelling, and finding blue objects in the residual images, which are otherwise blended in the foreground lens light profile. On the other hand, searching for gravitational lensing systems by identifying higher redshift emission lines hiding in the spectra of foreground galaxies have been proposed (Warren et al. Reference Warren, Hewett, Lewis, Møller, Iovino and Shaver1996; Hewett et al. Reference Hewett, Warren, Willis, Bland-Hawthorn and Lewis2000) and proved feasible (see e.g. Bolton et al. Reference Bolton, Treu, Koopmans, Gavazzi, Moustakas, Burles, Schlegel and Wayth2008b; Smith Reference Smith2017). Both detailed imaging and spectroscopic searches require extensive modelling and fitting. The colour and brightness selection of Belokurov et al. (Reference Belokurov2007) is easier to implement, yet it may deliver false-positives outnumbered the Einstein rings or arcs if the selection criteria are set too loose and require heavy human interactions (Lin et al. Reference Lin2009), or need further fine-tuning of the crietria (Belokurov et al. Reference Belokurov, Evans, Hewett, Moiseev, McMahon, Sanchez and King2009) and cannot be applied to other surveys.

To make better use of the colour and brightness selection method proposed by Belokurov et al. (Reference Belokurov2007) and to reduce human interactions during Einstein ring finding process, we propose to use a novel approach to identify circular patterns using computer vision techniques. We note that in this study, we concentrate on recognition of circular patterns that are applicable to Einstein rings and arcs. The drawback is that such algorithm cannot identify potential quad lenses, which are in the form of elliptical patterns.

This paper is organised as follows. In Section 2, we present the idea of using computer vision to find circles and arcs. We then apply this approach to several ongoing wide-field surveys and present our results in Section 3, followed by prospects in Section 4.

2 METHOD

Computer vision tackles tasks of reproducing the ability of human vision by electronically perceiving and understanding images, whose techniques can be easily and readily applied to astronomical data. One widely applied algorithm for computer vision to recognise parameterisable patterns is Hough transform (Hough Reference Hough1962; Duda & Hart Reference Duda and Hart1972), originally devised for identifying straight lines, but later generalised for any parameterisable patterns. The idea is, for a parameterisable shape, in our case a circle:

(1) $$\begin{equation} (x-a)^2 + (y-b)^2 = r^2, \end{equation}$$

the Hough transform constructs a quantised parameter space (dubbed ‘accumulator’) for the circle’s centre (a, b), and radius r. For every individual pixel on the image, the Hough transform then searches in the accumulator, and checks whether the pixels on the image can be described by a given set of parameters in the accumulator. This is considered as a voting process, and the set of parameters which obtain the highest numbers of votes, or pass a given threshold, will be regarded as a circle detection. For an n × n pixel image, Hough transform will construct an n 3 accumulator; hence, it will be computational expensive to blindly search for circles in large sky images. However, the Einstein rings have certain advantages to reduce the computational cost, as we will discuss in the next section.

3 APPLICATIONS

Hough transform has been used in astronomy for years, but mainly in removing satellite tracks (Cheselka Reference Cheselka1999) or detecting echelle orders (Ballester Reference Ballester1994). It has only been recently recognised to use Hough transform to detect extended circular and/or arc-like astronomical objects, such as supernova remnants, bent-tailed galaxies, radio relics, and so on, by Hollitt & Johnston-Hollitt (Reference Hollitt and Johnston-Hollitt2012). However, despite the promising results of detecting circular/arc features, Hollitt & Johnston-Hollitt (Reference Hollitt and Johnston-Hollitt2012) also pointed out that circle Hough transform will be computational expensive for blind search, and proposed a two-tired approach. i.e. first, find the location of possible extended circular objects, then perform Hough transform around this location, to reduce the computational efforts.

This being said, Einstein rings have several advantages to apply Hough transform. First of all, Einstein rings are mostly caused by massive objects, i.e. massive early type galaxies like LRGs, or galaxy clusters, which can be easily selected using their red colours. Second, the Einstein rings/arcs are relatively blue compared to the lens; hence, a handy criteria to pre-select Einstein ring candidates would be red objects associated with multiple blue objects nearby. Indeed, Belokurov et al. (Reference Belokurov2007) have shown that by selecting objects with r < 19.5 mag and gr > 0.6 mag associated with multiple faint (r > 19.5) and blue (gr < 0.5) sources within 6 arcsec, they can find the Cosmic Horseshoe and the 8 o’clock arc. However, as Lin et al. (Reference Lin2009) showed, even employing the aforementioned criteria, there are still thousands of candidates remain for visual inspection; hence, it is necessary to use circular pattern recognition algorithm as Hough transform mentioned here. Last but not the least, Einstein rings have relative small angular size (~ 5–10 arcsec in diameter). Since we know the range of the radius parameter space in priori, we can greatly reduce the computation time for Hough transform.

As a proof-of-concept, we use SDSS images to test the robustness of detecting Einstein rings and arcs using circle Hough transform. Since the main problem of automatic lens finding is to clean out false-positives, e.g. ring glaxies, face-on spirals, tidal tails, and chance alignments that appear like Einstein rings, we devise a two-tier process: (1) we pre-select LRGs associated with blue nearby objects; (2) we than feed these possible lens to circle Hough transform to identify Einstein rings and arcs.

To locate massive galaxies, we first start with the LRGs of Eisenstein et al. (Reference Eisenstein2001), which can be easily identified using colour selections. We then search for multiple (n ⩾ 2) blue (gr < 0.5) faint (r > 19.5 mag) objects associated with the LRGs within 6 arcsec. This reduce the number of possible lens LRGs from ~ 200 000 to 1 624, which are further passed to Hough transform to identify circles and arcs.

We use the circle Hough transform implementation from OpenCVFootnote 8 , an open source computer vision library. The program takes in an image of JPEG format to search for circular patterns. If the input is a colour image, the program will convert the input into an 8-bit, single-channel, gray-scale image. For convenience, we obtain the SDSS colour images from SDSS DR10 Finding Chart ToolFootnote 9 in batch mode, which can be tailored to retrieve images at a given position with different sizes and pixel scales. Since Einstein rings are relative small, we thus tailored our images to be 64 × 64 pixels, with pixel scale of 0.396 arcsec pixel−1 (the natural pixel scale of SDSS CCDs). We then feed the SDSS postage stamp images to the python version of OpenCV circle Hough transform. Since our images are rather small, the computational cost is negligible—it only takes 0.13 s to analyse each individual image on a 2.7 GHz Intel Core i5 processor. At the end, the Hough transform reduce the number of possible LRGs lens from 1 624 to 10, where four of them are spectroscopically confirmed wide arc lenses, suggesting a 100% completeness and 40% purity of our approach. Despite the small size of the image, and the fast calculation time, the circle patterns of Einsteinrings are well identified, even if they are only partial arcs. The results of our analysis are shown in Figure 1 and Table 1.

Figure 1. Applying Hough transform to Einstein rings from SDSS. Upper panels: images of (a) 8 o’clock (Allam et al. Reference Allam, Tucker, Lin, Diehl, Annis, Buckley-Geer and Frieman2007); (b) Cosmic Horseshoe (Belokurov et al. Reference Belokurov2007); (c) Clone (Lin et al. Reference Lin2009); (d) CSWA 7 (Belokurov et al. Reference Belokurov, Evans, Hewett, Moiseev, McMahon, Sanchez and King2009). Lower panels: Gray-scale images used for patterns recognition, and circular patterns identified by Hough transform (marked in green circles).

Table 1. Selection criteria.

While we show some success with the SDSS data, the sample size of lens systems is very small. In addition, the SDSS examples are all composed of wide-separation arcs. In this regard, we also perform testing using a larger sample from the CFHT legacy data. We make use of the public available images from SPACE WARPSFootnote 10 (Marshall et al. Reference Marshall2016; More et al. Reference More2016), which contains 59 high confident lens candidates identified by citizen scientists. Among the 59 SPACE WARPS lens systems, the circular Hough Transform is able to identify 53 lenses. The failed six cases are either arcs with small curvature or heavily blended lenses. In the first cases, because the arcs have very small curvature and resemble a line, the Hough transform failed to find a reasonable fit to a piece-wise circle/ring. In the later cases, because the arcs are heavily blended in the lens light, the Hough transform cannot distinguish the arcs from the lens light profile, thus failed to identify a circular pattern.

To test the versatility of our approach to different surveys, we also apply the Hough transform to Einstein rings discovered by DES (Canarias Einstein ring Bettinelli et al. Reference Bettinelli, Simioni, Aparicio, Hidalgo, Cassisi, Walker, Piotto and Valdes2016) and by HSC (Eye of Horus Tanaka et al. Reference Tanaka2016), both are serendipitous discoveries. The colour images are taken from the discovery papers. As shown in Figure 2, the circle Hough transform can identify both rings in DES and HSC without any ambiguity, demonstrating the versatility of our approach. Furthermore, since the lens of Eye of Horus is a BCG, it also shows that the Hough transform can work in crowded fields in galaxy clusters.

Figure 2. Applying Hough transform to Einstein rings from DECam, HSC, and UltraVISTA. Upper panels: images of (a) Canarias Einstein ring (from DES, Bettinelli et al. Reference Bettinelli, Simioni, Aparicio, Hidalgo, Cassisi, Walker, Piotto and Valdes2016); (b) Eye of Horus (from HSC, Tanaka et al. Reference Tanaka2016); (c) COSMOS 0050+4901 (from UltraVISTA, Hill et al. Reference Hill, Muzzin, Franx and van de Sande2016). Lower panels: Gray-scale images used for patterns recognition, and circular patterns identified by Hough transform (marked in green circles).

In addition, we also test the Hough transform on a quadruply lensed quiescent galaxy, COSMOS 0050+4901 (Hill et al. Reference Hill, Muzzin, Franx and van de Sande2016), discovered by the COSMOS/UltraVISTA survey. As the lensed images are relatively red, this demonstrates that the Hough transform can identify Einstein rings/arcs regardless of the source colours.

4 PROSPECTS

We have demonstrated the efficiency and robustness of finding Einstein rings, even partial arcs, in the ongoing large area surveys using the circle Hough transform. Though we only demonstrate its feasibility of LRG lensed Einstein rings, it can also be easily applied to Einstein rings/arcs from galaxy cluster lenses, assuming we know the positions of the galaxy clusters from BCG catalogues. In a more general context, such ring finder algorithms have wider applications in finding any circular patterns, for example, planetary nebulae, circumstellar shells around mass loss stars (AGB stars or LBVs), supernovae remnants, to name a few. The disadvantage of finding circular patterns around these astronomical objects, even if their locations can be inferred using certain selection criteria, is the large parameter space of their radii. The unknown radii will make traditional Hough transform computational expensive. To improve on this, there are some variants of circle Hough transform, e.g. probabilistic Hough transform and randomised Hough transform have been proposed. For example, Chiu, Lin, & Liaw (Reference Chiu, Lin and Liaw2010) have proposed a fast randomised Hough transform algorithm, which selects a random point on the image as seed point, and using a dedicated checking rule to confirm if the seed point is on a true circle or not. They have shown that such approach greatly reduces the computation effort, and is also suitable for noisy images.

In Section 3, we mentioned the computational requirements for Hough transform to go through each individual image, but not the total performance time, including pre-processing. In fact, the pre-processing will be the most time-consuming parts, including query of the candiate LRGs to obtain their coordinates, and subsequently making postage stamp images of these LRGs. In the case of SDSS, with its convenient SQL interface, we can easily query and obtain the coordinates of all of the candidate LRGs within a few hours. Given the coordinates, we can obtain the postage stamp in batch mode, which will cost less than an hour. As most of the ongoing and forthcoming all sky surveys (e.g. Pan-STARRS, HSC-SSPw) will also provide SQL interface for fast database query, as well as postage stamp service that can be fetched in batch mode, we expect that pre-processing time will be in the order of few hours.

In this study, we only discuss recognition of circular patterns. Nevertheless, it is possible to apply Hough transform for elliptical patterns. Such algorithms will be very useful in identifying Einstein cross, or quadruply lensed quasars (quads), where one pair of the four lensed quasars have larger projected distance to the lens galaxy than the other pair, and can be recognised as an ellipse. Quads are of great interests of the gravitational lensing community, not only because they can provide more constraints on the lens mass profile, but also due to their potential to determine Hubble constant using time-delay methods. With the advent of wide area sky surveys, Oguri & Marshall (Reference Oguri and Marshall2010) have predicted hundreds lensed quasars in HSC, thousands of lensed quasars in Pan-STARRS and DES, and ~ 10 000 of lensed quasars in the upcoming LSST, where 10–15% of them will be quads. In such wide area surveys, visual inspection to find quads is extremely difficult, if not impossible. Thus, efficient quads finding algorithm such as elliptical Hough transform will greatly reduce human interactions and provide an automated, robust method to identify interesting quads for the surveys to come.

ACKNOWLEDGEMENTS

Funding for the Sloan Digital Sky Survey IV has been provided by the Alfred P. Sloan Foundation, the U.S. Department of Energy Office of Science, and the Participating Institutions. SDSS-IV acknowledges support and resources from the Center for High-Performance Computing at the University of Utah. The SDSS website is www.sdss.org.

SDSS-IV is managed by the Astrophysical Research Consortium for the Participating Institutions of the SDSS Collaboration including the Brazilian Participation Group, the Carnegie Institution for Science, Carnegie Mellon University, the Chilean Participation Group, the French Participation Group, Harvard-Smithsonian Center for Astrophysics, Instituto de Astrofísica de Canarias, The Johns Hopkins University, Kavli Institute for the Physics and Mathematics of the Universe (IPMU)/University of Tokyo, Lawrence Berkeley National Laboratory, Leibniz Institut für Astrophysik Potsdam (AIP), Max-Planck-Institut für Astronomie (MPIA Heidelberg), Max-Planck-Institut für Astrophysik (MPA Garching), Max-Planck-Institut für Extraterrestrische Physik (MPE), National Astronomical Observatory of China, New Mexico State University, New York University, University of Notre Dame, Observatário Nacional/MCTI, The Ohio State University, Pennsylvania State University, Shanghai Astronomical Observatory, United Kingdom Participation Group, Universidad Nacional Autónoma de México, University of Arizona, University of Colorado Boulder, University of Oxford, University of Portsmouth, University of Utah, University of Virginia, University of Washington, University of Wisconsin, Vanderbilt University, and Yale University.

References

REFERENCES

Allam, S. S., Tucker, D. L., Lin, H., Diehl, H. T., Annis, J., Buckley-Geer, E. J., & Frieman, J. A. 2007, ApJ, 662, L51 CrossRefGoogle Scholar
Ballester, P. 1994, ASPC, 61, 319 Google Scholar
Belokurov, V., et al. 2007, ApJ, 671, L9 Google Scholar
Belokurov, V., Evans, N. W., Hewett, P. C., Moiseev, A., McMahon, R. G., Sanchez, S. F., & King, L. J. 2009, MNRAS, 392, 104 Google Scholar
Bettinelli, M., Simioni, M., Aparicio, A., Hidalgo, S. L., Cassisi, S., Walker, A. R., Piotto, G., & Valdes, F. 2016, MNRAS, 461, L67 Google Scholar
Bolton, A. S., Burles, S., Koopmans, L. V. E., Treu, T., Gavazzi, R., Moustakas, L. A., Wayth, R., & Schlegel, D. J. 2008a, ApJ, 682, 964984 Google Scholar
Bolton, A. S., Treu, T., Koopmans, L. V. E., Gavazzi, R., Moustakas, L. A., Burles, S., Schlegel, D. J., & Wayth, R. 2008b, ApJ, 684, 248259 Google Scholar
Brault, F., & Gavazzi, R. 2015, A&A, 577, A85 Google Scholar
Buckley-Geer, E. J., et al. 2011, ApJ, 742, 48 Google Scholar
Cabanac, R. A., Valls-Gabaud, D., Jaunsen, A. O., Lidman, C., & Jerjen, H. 2005, A&A, 436, L21 Google Scholar
Cheselka, M. 1999, ASPC, 172, 349 Google Scholar
Chiu, S. H., Lin, K. H., & Liaw, J. J. 2010, IJPRAI, 24, 457 Google Scholar
Chwolson, O. 1924, AN, 221, 329 Google Scholar
Collett, T. E. 2015, ApJ, 811, 20 Google Scholar
Duda, R. O., & Hart, P. E. 1972, Comm. ACM, 15, 11 Google Scholar
Einstein, A. 1915, SPAW, 844–847Google Scholar
Einstein, A. 1936, Science, 84, 506 Google Scholar
Eisenstein, D. J., et al. 2001, AJ, 122, 2267 CrossRefGoogle Scholar
Flaugher, B., et al. 2015, AJ, 150, 150 Google Scholar
Gavazzi, R., Marshall, P. J., Treu, T., & Sonnenfeld, A. 2014, ApJ, 785, 144 Google Scholar
Hewett, P. C., Warren, S. J., Willis, J. P., Bland-Hawthorn, J., & Lewis, G. F. 2000, ASPC, 195, 94 Google Scholar
Hill, A. R., Muzzin, A., Franx, M., & van de Sande, J. 2016, ApJ, 819, 74 Google Scholar
Hollitt, C., & Johnston-Hollitt, M. 2012, PASA, 29, 309 Google Scholar
Hough, P. V. C. 1962, U.S. Patent 3068654Google Scholar
Jones, T., Ellis, R. S., Richard, J., & Jullo, E. 2013, ApJ, 765, 48 CrossRefGoogle Scholar
Kaiser, N., et al. 2010, SPIE, 7733, 77330E Google Scholar
Kelly, P. L., et al. 2015, Science, 347, 1123 CrossRefGoogle Scholar
King, L. J., et al. 1998, MNRAS, 295, L41 Google Scholar
Kochanek, C. S., Keeton, C. R., & McLeod, B. A. 2001, ApJ, 547, 50 CrossRefGoogle Scholar
Lin, H., et al. 2009, ApJ, 699, 1242 CrossRefGoogle Scholar
LSST Science Collaboration et al. 2009, arXiv:0912.0201Google Scholar
Marshall, P. J., Hogg, D. W., Moustakas, L. A., Fassnacht, C. D., Bradač, M., Schrabback, T., & Blandford, R. D. 2009, ApJ, 694, 924 Google Scholar
Marshall, P. J., et al. 2016, MNRAS, 455, 1171 Google Scholar
Miyazaki, S., et al. 2006, SPIE, 6269, 62690B Google Scholar
More, A., et al. 2016, MNRAS, 455, 1191 Google Scholar
Narayan, R., & Bartelmann, M. 1996, astro, arXiv:astro-ph/9606001Google Scholar
Oguri, M., & Marshall, P. J. 2010, MNRAS, 405, 2579 Google Scholar
Paraficz, D., et al. 2016, A&A, 592, A75 Google Scholar
Refsdal, S. 1964, MNRAS, 128, 307 Google Scholar
Schmidt, B. P., Keller, S. C., Francis, P. J., & Bessell, M. S. 2005, AAS, 37, 15.09Google Scholar
Smith, R. J. 2017, MNRAS, 464, L46 Google Scholar
Stark, D. P., et al. 2013, MNRAS, 436, 1040 CrossRefGoogle Scholar
Tanaka, M., et al. 2016, ApJ, 826, L19 Google Scholar
Treu, T. 2010, ARA&A, 48, 87 Google Scholar
Vegetti, S., Lagattuta, D. J., McKean, J. P., Auger, M. W., Fassnacht, C. D., & Koopmans, L. V. E. 2012, Nature, 481, 341 Google Scholar
Warren, S. J., Hewett, P. C., Lewis, G. F., Møller, P., Iovino, A., & Shaver, P. A. 1996, MNRAS, 278, 139 Google Scholar
York, D. G., et al. 2000, AJ, 120, 1579 Google Scholar
Zheng, W., et al. 2012, Nature, 489, 406 CrossRefGoogle Scholar
Figure 0

Figure 1. Applying Hough transform to Einstein rings from SDSS. Upper panels: images of (a) 8 o’clock (Allam et al. 2007); (b) Cosmic Horseshoe (Belokurov et al. 2007); (c) Clone (Lin et al. 2009); (d) CSWA 7 (Belokurov et al. 2009). Lower panels: Gray-scale images used for patterns recognition, and circular patterns identified by Hough transform (marked in green circles).

Figure 1

Table 1. Selection criteria.

Figure 2

Figure 2. Applying Hough transform to Einstein rings from DECam, HSC, and UltraVISTA. Upper panels: images of (a) Canarias Einstein ring (from DES, Bettinelli et al. 2016); (b) Eye of Horus (from HSC, Tanaka et al. 2016); (c) COSMOS 0050+4901 (from UltraVISTA, Hill et al. 2016). Lower panels: Gray-scale images used for patterns recognition, and circular patterns identified by Hough transform (marked in green circles).