diff --git a/648.js b/648.js index 4b166ac..31005b0 100644 --- a/648.js +++ b/648.js @@ -1 +1 @@ -(self.webpackChunkrlraudioweb=self.webpackChunkrlraudioweb||[]).push([[648],{798:(e,t,n)=>{var r={"./Database.json":932};function a(e){var t=o(e);return n(t)}function o(e){if(!n.o(r,e)){var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}return r[e]}a.keys=function(){return Object.keys(r)},a.resolve=o,e.exports=a,a.id=798},998:(e,t,n)=>{var r={"./HRTF_Database_AVAR2024_v5.pdf":693,"./RLR Audio Web.pdf":492};function a(e){var t=o(e);return n(t)}function o(e){if(!n.o(r,e)){var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}return r[e]}a.keys=function(){return Object.keys(r)},a.resolve=o,e.exports=a,a.id=998},311:(e,t,n)=>{var r={"./HRIRs.zip":401,"./HRIRs_Repeated_Measurements.zip":254,"./HRIRs_mannequins.zip":889,"./Hp_Filters.zip":692,"./SS2_HRTF_FullDatabase.zip":85};function a(e){var t=o(e);return n(t)}function o(e){if(!n.o(r,e)){var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}return r[e]}a.keys=function(){return Object.keys(r)},a.resolve=o,e.exports=a,a.id=311},648:(e,t,n)=>{"use strict";n.r(t),n.d(t,{default:()=>Y});var r=n(540),a=n(485),o=n.n(a),i=n(228),l={img:function(e){return r.createElement("span",{className:"w-full flex flex-col items-center"},r.createElement("img",e))}};const s=function(e){var t=e.children;return r.createElement(i.$,{components:l},t)};var c=["mb-7","prose","lg:prose-lg","xl:prose-xl","2xl:prose-xl","dark:prose-invert","sm:prose-sm"];const u=function(e){var t=e.children;return r.createElement("article",{className:o()(c)},t)},m=function(){return r.createElement(u,null,r.createElement(s,null,"## Sound Sphere 2: A High-resolution HRTF Database\n\nThe RLR HRTF Database is a high-resolution database of 78 head-related transfer functions (HRTFs) that were collected during the 2022 Audio Engineering Society AVAR conference. The database is freely available for download.\n"))},f=function(){return r.createElement(u,null,r.createElement(s,null,"### Database details\n\n\nHRTFs were collected in a purpose-built anechoic chamber housing a vertically oriented, motorized arc with a 2-meter-radius, which contains 54 loudspeakers spaced every 3° in elevation. Participants were aligned using a height-adjustable platform and wall-mounted cross lasers; their head position and orientation were tracked in real-time during the measurement. Inter- and intra-participant error analyses across more than 1600 incident angles per participant indicate high precision in measurements. In addition to being a high-resolution and a high-precision HRTF database, this database includes the corresponding equalization filters for both commercial headphones and a VR headset, as well as some demographic information for each participant. \n\n![image tag](FRL-Research-Audio-Anechoic-Chamber-2.webp)\n\nA relatively large number of the participants in this database are researchers from the spatial audio research community (who participated in the AVAR 2022 conference). The authors hope that publishing this database may facilitate collaborations in the future.\n\nFor more details, please see the [academic publication (link to be updated)](https://www.scholear.google.com).\n\nThe Audio Team at Reality Labs Research (RLR Audio) is working on novel technologies to enable both audio presence and augmented hearing, letting us hear better in noisy environments with our future AR glasses. In our [blog post](https://about.fb.com/news/2020/09/facebook-reality-labs-research-future-of-audio/) we take you behind the scenes with the team for an in-depth look at some of our demos and prototype technology. \n\n\n### Citation\nPlease note that use of the freely available database should go along with citing of the publication above. \n\n```\n@article{warneckeHRTF2024,\n title = {Sound Sphere 2: A High-resolution HRTF Database,\n author = {Michaela Warnecke, Samuel Clapp, Zamir Ben-Hur, David Lou Alon, Sebastià V. Amengual Garí, and Paul Calamia},\n booktitle = {2024 AES 5th International Conference on Audio for Virtual and Augmented Reality},\n doi = {}\n month = {8}\n year = {2024}\n}\n```\n\n\n---\n\n### License\n\nThe dataset is distributed under the CC-BY-4.0 license, as found in the [LICENSE](https://github.com/facebookresearch/SS2_HRTF/blob/main/LICENSE) file.\n\n### Additional References\n\nFor more work by the RLR Audio Team, please see a selection of recent related publications below: \n\n\n#### 2024\n- [HRTF Notch Prediction using Anthropometric Features and Neural Networks](https://ieeexplore.ieee.org/abstract/document/10447610) Arbel et al., 2024\n\n#### 2022\n- [Ear shape-based HRTF personalization.](https://www.aes.org/e-lib/browse.cfm?elib=21873) Warnecke et al., 2022\n- [Remote testing](https://asa.scitation.org/doi/full/10.1121/10.0010422) Peng et al., 2022\n- [HATS digital twin](http://www.aes.org/e-lib/browse.cfm?elib=21855) Hajarolasvadi et al., 2022\n- [Verification on head-related transfer functions of a rigid sphere model](https://doi.org/10.1121/10.0011736) Meyer et al., 2022\n\n#### 2021\n- [HRTF metrics for localization perception.](https://asa.scitation.org/doi/full/10.1121/10.0003983) Ananthabhotla et al., 2021\n- [DNN HRTF prediction from ear shapes](https://research.facebook.com/publications/on-the-predictability-of-hrtfs-from-ear-shapes-using-deep-networks/) Zhou et al., 2021\n- [Sparse HRTF sampling](https://research.facebook.com/publications/sparse-head-related-transfer-function-representation-with-spatial-aliasing-cancellation/) Arbel et al., 2021\n- [Omnidirectional source for near-field HRTF measurements](https://research.facebook.com/publications/full-range-omnidirectional-sound-source-for-near-field-head-related-transfer-functions-measurement/) Chojnacki et al., 2021\n\n#### 2020\n- [Numerical simulations of near-field HRTF](https://research.facebook.com/publications/numerical-simulations-of-near-field-head-related-transfer-functions-magnitude-verification-and-validation-with-laser-spark-sources/) Prepeliţă et al., 2020\n- [FDTD simulations of pinna-related TFs](https://research.facebook.com/publications/pinna-related-transfer-functions-and-lossless-wave-equation-using-finite-difference-methods-validation-with-measurements/) Prepeliţă et al., 2020\n\n#### 2019\n- [Sparse HRTF representation with ear alignment](https://research.facebook.com/publications/efficient-representation-and-sparse-sampling-of-head-related-transfer-functions-using-phase-correction-based-on-ear-alignment/) Ben-Hur et al., 2019\n\n\n\n\n\n"))};function d(e){for(var t=0;e/1e3>1;)t++,e/=1e3;return"".concat(Math.round(e)).concat(["bytes","kb","mb","gb"][t])}var p=["prose","dark:prose-invert"],h=function(e){var t=e.children;return r.createElement("span",{className:a(p)}," ",r.createElement("h3",null,t)," ")},b=function(e){var t=e.children;return r.createElement("span",{className:a(p)}," ",r.createElement("h4",null,t)," ")},y=function(e){var t=e.children;return r.createElement("span",{className:a(p)}," ",r.createElement("h5",null,t)," ")},g=["bg-gray-300","hover:bg-gray-400","text-gray-800","font-bold","py-2","px-4","rounded","inline-flex","items-center"],v=["fill-current","w-4","h-4","mr-2"];const w=function(e){var t=e.url;return r.createElement("a",{href:t},r.createElement("button",{className:o()(g)},r.createElement("svg",{className:o()(v),xmlns:"http://www.w3.org/2000/svg",viewBox:"0 0 20 20"},r.createElement("path",{d:"M13 8V2H7v6H2l8 8 8-8h-5zM0 18h20v2H0v-2z"})),r.createElement("span",null,"Download")))};function R(e){return function(e){var t=e.split("://");if(t.length>1)return!0;var n=t[0].split("/");return n.length>1&&n[0].includes(".")}(e)?e:("/"==e[0]&&(e=e.substr(1,e.length-1)),"".concat(location.href).concat(e))}function E(e,t){(null==t||t>e.length)&&(t=e.length);for(var n=0,r=new Array(t);ne.length)&&(t=e.length);for(var n=0,r=new Array(t);ne.length)&&(t=e.length);for(var n=0,r=new Array(t);ne.length)&&(t=e.length);for(var n=0,r=new Array(t);n{"use strict";e.exports=n.p+"HRIRs.zip"},254:(e,t,n)=>{"use strict";e.exports=n.p+"HRIRs_Repeated_Measurements.zip"},889:(e,t,n)=>{"use strict";e.exports=n.p+"HRIRs_mannequins.zip"},693:(e,t,n)=>{"use strict";e.exports=n.p+"HRTF_Database_AVAR2024_v5.pdf"},692:(e,t,n)=>{"use strict";e.exports=n.p+"Hp_Filters.zip"},492:(e,t,n)=>{"use strict";e.exports=n.p+"RLR Audio Web.pdf"},85:(e,t,n)=>{"use strict";e.exports=n.p+"SS2_HRTF_FullDatabase.zip"},932:e=>{"use strict";e.exports=JSON.parse('{"title":"HRTF Database","item":{"name":"Full Dataset","url":"https://github.com/facebook/rlraudioweb/raw/main/files/SS2_HRTF_FullDatabase.zip?download="},"subitems":[{"name":"HRIRs (human participants)","url":"https://github.com/facebook/rlraudioweb/raw/main/files/HRIRs.zip?download="},{"name":"HRIRs (mannequins)","url":"https://github.com/facebook/rlraudioweb/raw/main/files/HRIRs_mannequins.zip?download="},{"name":"HRIRs (repeated human participants)","url":"https://github.com/facebook/rlraudioweb/raw/main/files/HRIRs_Repeated_Measurements.zip?download="},{"name":"Headphone Filters","url":"https://github.com/facebook/rlraudioweb/raw/main/files/Hp_Filters.zip?download="},{"name":"Demographic Data","url":"https://github.com/facebook/rlraudioweb/raw/main/files/Demographics.xls?download="},{"name":"Docmmentation","url":"/HRTF_Database_AVAR2024_v5.pdf"}]}')}}]); \ No newline at end of file +(self.webpackChunkrlraudioweb=self.webpackChunkrlraudioweb||[]).push([[648],{798:(e,t,n)=>{var r={"./Database.json":932};function a(e){var t=o(e);return n(t)}function o(e){if(!n.o(r,e)){var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}return r[e]}a.keys=function(){return Object.keys(r)},a.resolve=o,e.exports=a,a.id=798},998:(e,t,n)=>{var r={"./HRTF_Database_AVAR2024_v5.pdf":693,"./RLR Audio Web.pdf":492};function a(e){var t=o(e);return n(t)}function o(e){if(!n.o(r,e)){var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}return r[e]}a.keys=function(){return Object.keys(r)},a.resolve=o,e.exports=a,a.id=998},311:(e,t,n)=>{var r={"./HRIRs.zip":401,"./HRIRs_Repeated_Measurements.zip":254,"./HRIRs_mannequins.zip":889,"./Hp_Filters.zip":692,"./SS2_HRTF_FullDatabase.zip":85};function a(e){var t=o(e);return n(t)}function o(e){if(!n.o(r,e)){var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}return r[e]}a.keys=function(){return Object.keys(r)},a.resolve=o,e.exports=a,a.id=311},648:(e,t,n)=>{"use strict";n.r(t),n.d(t,{default:()=>Y});var r=n(540),a=n(485),o=n.n(a),i=n(228),l={img:function(e){return r.createElement("span",{className:"w-full flex flex-col items-center"},r.createElement("img",e))}};const s=function(e){var t=e.children;return r.createElement(i.$,{components:l},t)};var c=["mb-7","prose","lg:prose-lg","xl:prose-xl","2xl:prose-xl","dark:prose-invert","sm:prose-sm"];const u=function(e){var t=e.children;return r.createElement("article",{className:o()(c)},t)},m=function(){return r.createElement(u,null,r.createElement(s,null,"## Sound Sphere 2: A High-resolution HRTF Database\n\nThe RLR HRTF Database is a high-resolution database of 78 head-related transfer functions (HRTFs) that were collected during the 2022 Audio Engineering Society AVAR conference. The database is freely available for download.\n"))},f=function(){return r.createElement(u,null,r.createElement(s,null,"### Database details\n\n\nHRTFs were collected in a purpose-built anechoic chamber housing a vertically oriented, motorized arc with a 2-meter-radius, which contains 54 loudspeakers spaced every 3° in elevation. Participants were aligned using a height-adjustable platform and wall-mounted cross lasers; their head position and orientation were tracked in real-time during the measurement. Inter- and intra-participant error analyses across more than 1600 incident angles per participant indicate high precision in measurements. In addition to being a high-resolution and a high-precision HRTF database, this database includes the corresponding equalization filters for both commercial headphones and a VR headset, as well as some demographic information for each participant. \n\n![image tag](FRL-Research-Audio-Anechoic-Chamber-2.webp)\n\nA relatively large number of the participants in this database are researchers from the spatial audio research community (who participated in the AVAR 2022 conference). The authors hope that publishing this database may facilitate collaborations in the future.\n\nFor more details, please see the [academic publication (link to be updated)](https://www.scholear.google.com).\n\nThe Audio Team at Reality Labs Research (RLR Audio) is working on novel technologies to enable both audio presence and augmented hearing, letting us hear better in noisy environments with our future AR glasses. In our [blog post](https://about.fb.com/news/2020/09/facebook-reality-labs-research-future-of-audio/) we take you behind the scenes with the team for an in-depth look at some of our demos and prototype technology. \n\n\n### Citation\nPlease note that use of the freely available database should go along with citing of the publication above. \n\n```\n@article{warneckeHRTF2024,\n title = {Sound Sphere 2: A High-resolution HRTF Database,\n author = {Michaela Warnecke, Samuel Clapp, Zamir Ben-Hur, David Lou Alon, Sebastià V. Amengual Garí, and Paul Calamia},\n booktitle = {2024 AES 5th International Conference on Audio for Virtual and Augmented Reality},\n doi = {}\n month = {8}\n year = {2024}\n}\n```\n\n\n---\n\n### License\n\nThe dataset is distributed under the CC-BY-4.0 license, as found in the [LICENSE](https://github.com/facebookresearch/SS2_HRTF/blob/main/LICENSE) file.\n\n### Additional References\n\nFor more work by the RLR Audio Team, please see a selection of recent related publications below: \n\n\n#### 2024\n- [HRTF Notch Prediction using Anthropometric Features and Neural Networks](https://ieeexplore.ieee.org/abstract/document/10447610) Arbel et al., 2024\n\n#### 2022\n- [Ear shape-based HRTF personalization.](https://www.aes.org/e-lib/browse.cfm?elib=21873) Warnecke et al., 2022\n- [Remote testing](https://asa.scitation.org/doi/full/10.1121/10.0010422) Peng et al., 2022\n- [HATS digital twin](http://www.aes.org/e-lib/browse.cfm?elib=21855) Hajarolasvadi et al., 2022\n- [Verification on head-related transfer functions of a rigid sphere model](https://doi.org/10.1121/10.0011736) Meyer et al., 2022\n\n#### 2021\n- [HRTF metrics for localization perception.](https://asa.scitation.org/doi/full/10.1121/10.0003983) Ananthabhotla et al., 2021\n- [DNN HRTF prediction from ear shapes](https://research.facebook.com/publications/on-the-predictability-of-hrtfs-from-ear-shapes-using-deep-networks/) Zhou et al., 2021\n- [Sparse HRTF sampling](https://research.facebook.com/publications/sparse-head-related-transfer-function-representation-with-spatial-aliasing-cancellation/) Arbel et al., 2021\n- [Omnidirectional source for near-field HRTF measurements](https://research.facebook.com/publications/full-range-omnidirectional-sound-source-for-near-field-head-related-transfer-functions-measurement/) Chojnacki et al., 2021\n\n#### 2020\n- [Numerical simulations of near-field HRTF](https://research.facebook.com/publications/numerical-simulations-of-near-field-head-related-transfer-functions-magnitude-verification-and-validation-with-laser-spark-sources/) Prepeliţă et al., 2020\n- [FDTD simulations of pinna-related TFs](https://research.facebook.com/publications/pinna-related-transfer-functions-and-lossless-wave-equation-using-finite-difference-methods-validation-with-measurements/) Prepeliţă et al., 2020\n\n#### 2019\n- [Sparse HRTF representation with ear alignment](https://research.facebook.com/publications/efficient-representation-and-sparse-sampling-of-head-related-transfer-functions-using-phase-correction-based-on-ear-alignment/) Ben-Hur et al., 2019\n\n\n\n\n\n"))};function d(e){for(var t=0;e/1e3>1;)t++,e/=1e3;return"".concat(Math.round(e)).concat(["bytes","kb","mb","gb"][t])}var p=["prose","dark:prose-invert"],h=function(e){var t=e.children;return r.createElement("span",{className:a(p)}," ",r.createElement("h3",null,t)," ")},b=function(e){var t=e.children;return r.createElement("span",{className:a(p)}," ",r.createElement("h4",null,t)," ")},y=function(e){var t=e.children;return r.createElement("span",{className:a(p)}," ",r.createElement("h5",null,t)," ")},g=["bg-gray-300","hover:bg-gray-400","text-gray-800","font-bold","py-2","px-4","rounded","inline-flex","items-center"],v=["fill-current","w-4","h-4","mr-2"];const w=function(e){var t=e.url;return r.createElement("a",{href:t},r.createElement("button",{className:o()(g)},r.createElement("svg",{className:o()(v),xmlns:"http://www.w3.org/2000/svg",viewBox:"0 0 20 20"},r.createElement("path",{d:"M13 8V2H7v6H2l8 8 8-8h-5zM0 18h20v2H0v-2z"})),r.createElement("span",null,"Download")))};function R(e){return function(e){var t=e.split("://");if(t.length>1)return!0;var n=t[0].split("/");return n.length>1&&n[0].includes(".")}(e)?e:("/"==e[0]&&(e=e.substr(1,e.length-1)),"".concat(location.href).concat(e))}function E(e,t){(null==t||t>e.length)&&(t=e.length);for(var n=0,r=new Array(t);ne.length)&&(t=e.length);for(var n=0,r=new Array(t);ne.length)&&(t=e.length);for(var n=0,r=new Array(t);ne.length)&&(t=e.length);for(var n=0,r=new Array(t);n{"use strict";e.exports=n.p+"HRIRs.zip"},254:(e,t,n)=>{"use strict";e.exports=n.p+"HRIRs_Repeated_Measurements.zip"},889:(e,t,n)=>{"use strict";e.exports=n.p+"HRIRs_mannequins.zip"},693:(e,t,n)=>{"use strict";e.exports=n.p+"HRTF_Database_AVAR2024_v5.pdf"},692:(e,t,n)=>{"use strict";e.exports=n.p+"Hp_Filters.zip"},492:(e,t,n)=>{"use strict";e.exports=n.p+"RLR Audio Web.pdf"},85:(e,t,n)=>{"use strict";e.exports=n.p+"SS2_HRTF_FullDatabase.zip"},932:e=>{"use strict";e.exports=JSON.parse('{"title":"HRTF Database","item":{"name":"Full Dataset","url":"https://github.com/facebook/rlraudioweb/raw/main/files/SS2_HRTF_FullDatabase.zip?download="},"subitems":[{"name":"HRIRs (human participants)","url":"https://github.com/facebook/rlraudioweb/raw/main/files/HRIRs.zip?download="},{"name":"HRIRs (mannequins)","url":"https://github.com/facebook/rlraudioweb/raw/main/files/HRIRs_mannequins.zip?download="},{"name":"HRIRs (repeated human participants)","url":"https://github.com/facebook/rlraudioweb/raw/main/files/HRIRs_Repeated_Measurements.zip?download="},{"name":"Headphone Filters","url":"https://github.com/facebook/rlraudioweb/raw/main/files/Hp_Filters.zip?download="},{"name":"Demographic Data","url":"https://github.com/facebook/rlraudioweb/raw/main/files/Demographics.xls?download="},{"name":"Documentation","url":"/HRTF_Database_AVAR2024_v5.pdf"}]}')}}]); \ No newline at end of file